[lustre-discuss] set OSTs read only ?
Bob Ball
ball at umich.edu
Wed Jul 12 19:49:46 PDT 2017
On the mgs/mgt do something like:
lctl --device <fsname>-OST0019-osc-MDT0000 deactivate
No further files will be assigned to that OST. Reverse with
"activate". Or reboot the mgs/mdt as this is not persistent. "lctl dl"
will tell you exactly what that device name should be for you.
bob
On 7/12/2017 6:04 PM, Alexander I Kulyavtsev wrote:
> You may find advise from Andreas on this list (also attached below). I
> did not try setting fail_loc myself.
>
> In 2.9 there is setting osp.*.max_create_count=0described at LUDOC-305.
>
> We used to set OST degraded as described in lustre manual.
> It works most of the time but at some point I saw lustre errors in
> logs for some ops. Sorry, I do not recall details.
>
> I still not sure either of these approaches will work for you: setting
> OST degraded or fail_loc will makes some osts selected instead of others.
> You may want to verify if these settings will trigger clean error on
> user side (instead of blocking) when all OSTs are degraded.
>
> The other and also simpler approach would be to enable lustre quota
> and set quota below used space for all users (or groups).
>
> Alex.
>
>> *From: *"Dilger, Andreas" <andreas.dilger at intel.com
>> <mailto:andreas.dilger at intel.com>>
>> *Subject: **Re: [lustre-discuss] lustre 2.5.3 ost not draining*
>> *Date: *July 28, 2015 at 11:51:38 PM CDT
>> *Cc: *"lustre-discuss at lists.lustre.org
>> <mailto:lustre-discuss at lists.lustre.org>"
>> <lustre-discuss at lists.lustre.org
>> <mailto:lustre-discuss at lists.lustre.org>>
>>
>> Setting it degraded means the MDS will avoid allocations on that OST
>> unless there aren't enough OSTs to meet the request (e.g. stripe_count =
>> -1), so it should work.
>>
>> That is actually a very interesting workaround for this problem, and it
>> will work for older versions of Lustre as well. It doesn't disable the
>> OST completely, which is fine if you are doing space balancing (and may
>> even be desirable to allow apps that need more bandwidth for a widely
>> striped file), but it isn't good if you are trying to empty the OST
>> completely to remove it.
>>
>> It looks like another approach would be to mark the OST as having no free
>> space using OBD_FAIL_OST_ENOINO (0x229) fault injection on that OST:
>>
>> lctl set_param fail_loc=0x229 fail_val=<ost_index>
>>
>> This would cause the OST to return 0 free inodes from OST_STATFS for the
>> specified OST index, and the MDT would skip this OST completely. To
>> disable all of the OSTs on an OSS use <ost_index> = -1. It isn't
>> possible
>> to selectively disable a subset of OSTs using this method. The
>> OBD_FAIL_OST_ENOINO fail_loc has been available since Lustre 2.2, which
>> covers all of the 2.4+ versions that are affected by this issue.
>>
>> If this mechanism works for you (it should, as this fail_loc is used
>> during regular testing) I'd be obliged if someone could file an LUDOC bug
>> so the manual can be updated.
>>
>> Cheers, Andreas
>
>
>> On Jul 12, 2017, at 4:20 PM, Riccardo Veraldi
>> <Riccardo.Veraldi at cnaf.infn.it
>> <mailto:Riccardo.Veraldi at cnaf.infn.it>> wrote:
>>
>> Hello,
>>
>> on one of my lustre FS I need to find a solution so that users can still
>> access data on the FS but cannot write new files on it.
>> I have hundreds of clients accessing the FS so remounting it ro is not
>> really easily feasible.
>> Is there an option on the OSS side to allow OSTs to be accessed just to
>> read data and not to store new data ?
>> tunefs.lustre could do that ?
>> thank you
>>
>> Rick
>>
>> _______________________________________________
>> lustre-discuss mailing list
>> lustre-discuss at lists.lustre.org <mailto:lustre-discuss at lists.lustre.org>
>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>
>
>
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss at lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20170712/ab93bb4d/attachment.htm>
More information about the lustre-discuss
mailing list