[lustre-discuss] lustre-discuss Digest, Vol 122, Issue 9
Dilger, Andreas
andreas.dilger at intel.com
Fri May 6 15:02:31 PDT 2016
On 2016/05/06, 15:48, "lustre-discuss on behalf of Fernando Pérez"
<lustre-discuss-bounces at lists.lustre.org on behalf of fperez at icm.csic.es>
wrote:
>Thank you Mark.
>
>Finally I have killed the e2fsck. After restart again our lustre
>filesystem it seems all works OK.
>
>We are using two 300 GB RAID 1 10K SAS drives for the combined mdt / mgs.
>
>I tried to run the e2fsck -fy because the -fn finish in 2 hoursŠI think
>there is a problem in the latest e2fsprogs because the e2fsck returned
>that it was repairing more inodes than our filesystem has.
Which specific version of e2fsprogs are you using?
Cheers, Andreas
>
>Regards.
>=============================================
>Fernando Pérez
>Institut de Ciències del Mar (CMIMA-CSIC)
>Departament Oceanografía Física i Tecnològica
>Passeig Marítim de la Barceloneta,37-49
>08003 Barcelona
>Phone: (+34) 93 230 96 35
>=============================================
>
>> El 6 may 2016, a las 17:57, Mark Hahn <hahn at mcmaster.ca> escribió:
>>
>>> More information about our lustre system: combined mds / mdt has 189
>>>GB and
>>> 8.9 GB used. It was formatted with the default options.
>>
>> fsck time is more about the number of files (inodes), rather than
>> the size. but either you have quite slow storage, or something is
>>wrong.
>>
>> as a comparison point, I can do a full/force fsck on one of our MDS/MDT
>> that has 143G or 3.3T in use (313M inodes) in about 2 hours. it is a MD
>> raid10 on 16x 10K SAS drives, admittedly.
>>
>> if your hardware is conventional (locally-attached multi-disk RAID),
>> it might make sense to look at its configuration. for instance, fsck
>> is largely seek-limited, but doing too much readahead, or using large
>> RAID block sizes (for R5/6) can be disadvantageous. having plenty of
>>RAM helps in some phases.
>>
>> regards, mark hahn.
>
>_______________________________________________
>lustre-discuss mailing list
>lustre-discuss at lists.lustre.org
>http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>
Cheers, Andreas
--
Andreas Dilger
Lustre Principal Architect
Intel High Performance Data Division
More information about the lustre-discuss
mailing list