[Lustre-discuss] is Luster ready for prime time?

Lind, Bobbie J bobbie.j.lind at intel.com
Mon Jan 21 11:25:17 PST 2013


Indivar,

I would be very interested to see what tuning parameters you have set to
tune lustre and the storage for small files.  I have had similar setups in
the past and been stumped by the small file performance.

-- 
Bobbie Lind



>Date: Mon, 21 Jan 2013 11:24:32 -0500
>From: greg whynott <greg.whynott at gmail.com>
>Subject: Re: [Lustre-discuss] is Luster ready for prime time?
>To: Indivar Nair <indivar.nair at techterra.in>
>Cc: "lustre-discuss at lists.lustre.org"
>	<lustre-discuss at lists.lustre.org>
>Message-ID:
>	<CAKuzA1G4-W122LQrf3VKqADd=WrDgcAVx5hyAGJfZwwR8KKG2g at mail.gmail.com>
>Content-Type: text/plain; charset="utf-8"
>
>Thanks very much Indivar,  informative read.    it is good to see others
>in
>our sector are using the technology and you have some good points.
>
>have a great day,
>greg
>
>
>
>On Sat, Jan 19, 2013 at 6:52 AM, Indivar Nair
><indivar.nair at techterra.in>wrote:
>
>>  Hi Greg,
>>
>> One of our customers had a similar requirement and we deployed Lustre
>> 2.0.0.1 for them. This was in July 2011. Though there were a lots of
>> problems initially, all of them were sorted out over time. They are
>>quite
>> happy with it now.
>>
>> *Environment:*
>> Its a 150 Artist studio with around 60 Render nodes. The studio mainly
>> uses Mocha, After Effects, Silhouette, Synth Eye, Maya, and Nuke among
>> others. They mainly work on 3D Effects and Stereoscopy Conversions.
>> Around 45% of Artists and Render Nodes are on Linux and use native
>>Lustre
>> Client. All others access it through Samba.
>>
>> *Lustre Setup:*
>> It consists of 2 x Dell R610 as MDS Nodes, and 4 x Dell R710 as OSS
>>Nodes.
>> 2 x Dell MD3200 with 12x1TB SAS Nearline Disks are used for storage.
>>Each
>> Dell MD3200s are shared among 2 OSS nodes for H/A.
>>
>> Since the original plan (which didn't happen) was to move to a 100%
>>Linux
>> environment, we didn't allocate separate Samba Gateways and use the OSS
>> nodes with CTDB for it. Thankfully, we haven't had any issues with that
>>yet.
>>
>> *Performance:*
>> We get a good THROUGHPUT of 800 - 1000MB/s with Lustre Caching. The
>>disks
>> it self provide much lesser speeds. But that is fine, as caching is in
>> effect most of the time.
>>
>> *Challenge:*
>> The challenge for us was to tune the storage for small files 10 - 50MB
>> totalling to 10s of GBs. An average shot would consist of 2000 - 4000
>>.dpx
>> images. Some Scenes / Shots also had millions of <1MB Maya Cache files.
>> This did tax the storage, especially the MDS. Fixed it to an extent by
>> adding more RAM to MDS.
>>
>> *Suggestions:*
>>
>> 1. Get the real number of small files (I mean <1MB ones) created / used
>>by
>> all software. These are the ones that could give you the most trouble.
>>Do
>> not assume anything.
>>
>> 2. Get the file - sizes, numbers and access patterns absolutely correct.
>> This is the key.
>>     Its easier to design and tune Lustre for large files and I/O.
>>
>> 3. Network tuning is as important and storage tuning. Tune Switches,
>>each
>> Workstation, Render Nodes, Samba / NFS Gateways, OSS Nodes, MDS Nodes,
>> everything.
>>
>> 4. Similarly do not undermine Samba / NFS Gateway. Size and tune them
>> correctly too.
>>
>> 5. Use High Speed Switching like QDR Infiniband or 40GigE, especially
>>for
>> backend connectivity between Samba/NFS Gateway and Lustre MDS/OSS Nodes.
>>
>> 6. As far as possible, have fixed directory pattern for all projects.
>> Separate working files (Maya, Nuke, etc.) from the data, i.e. frames /
>> images, videos, etc. at the top directory level it self. This will help
>>you
>> tune / manage the storage better. Different directory tree for different
>> file sizes or file access types.
>>
>> If designed and tuned right, I think Lustre is best storage currently
>> available for your kind of work.
>>
>> Hope this helps.
>>
>> Regards,
>>
>>
>> Indivar Nair
>>
>>
>> On Fri, Jan 18, 2013 at 1:51 AM, greg whynott
>><greg.whynott at gmail.com>wrote:
>>
>>> Hi Charles,
>>>
>>>   I received a few off list challenging email messages along with a few
>>> fishing ones,  but its all good.   its interesting how a post asking a
>>> question can make someone appear angry.  8)
>>>
>>> Our IO profiles from the different segments of our business do vary
>>> greatly.   The HPC is more or less the typical load you would expect to
>>> see,  depending on which software is in use for the for the job being
>>>ran.
>>>       We have hundreds of artists and administrative staff who use the
>>>file
>>> system in a variety of ways.   Some examples would include but not
>>>limited
>>> to:  saving out multiple revisions of photoshop documents (typically
>>>in the
>>> hundreds of megs to +1gig range),   video editing (stereoscopic 2k and
>>>4k
>>> images(again from 10's 100's to gigs in size) including uncompressed
>>> video,  excel, word and similar files,  thousands of project files
>>>(from
>>> software such as Maya,  Nuke and similar)  these also vary largely in
>>>size,
>>> from 1 to thousands of megs in size.
>>>
>>> The intention is keep our data bases and VM requirements on the
>>>existing
>>> file system which is comprised of about 100 10k SAS drives,  it works
>>>well.
>>>
>>> We did consider GPFS but that consideration went out the door once I
>>> started talking to them and hammering in some numbers into their online
>>> calculator.  Things got a bit crazy quickly.   They have different
>>>pricing
>>> for the different types and speeds of Intel CPUs.  I got the feeling
>>>they
>>> were trying to squeeze every penny out of customers they could.  felt
>>>very
>>> Brocade-ish and left a bad taste with us.   wouldn't of been much of a
>>> problem as some other shops I've worked at,  but here we do have a
>>>finite
>>> budget to work within.
>>>
>>> The NAS vendors could all be considered scale out I suspect.   All 3
>>>can
>>> scale out the storage and front end.  NA C-mode can have up to 24
>>>heads,
>>> Blue Arc goes up to 4 or 8 depending on the class,  Isilon can go up
>>>to 24
>>> nodes or more as well if memory serves me correctly,  and they all
>>>have a
>>> single name space solution in place.   They each have their limits,
>>>but
>>> for our use case they are really subjective.   We will not hit the
>>>limits
>>> of their scalability before we are considering a fork lift refresh.
>>>In our
>>> view,  for what they offer it is perty much a wash for them - any would
>>> meet our needs.  NetApp still has a silly agg/vol size limit,  at
>>>least it
>>> is up to 90TB now (from 9 in the past(formatted fs use))..  in April
>>>it is
>>> suppose to go much higher.
>>>
>>>  The block storage idea in the mix - since all our HPC is linux,  they
>>> all would become luster clients.   To provide a gateway into the luster
>>> storage for none linux/luster hosts the thinking was a clustered pair
>>>of
>>> linux boxes running SAMBA/NFS which were also Luster clients.    Its
>>>just
>>> an idea being bounced around at this point.  The data serving
>>>requirements
>>> of the non HPC parts of the business are much less.   The video editors
>>> most likely would stay on our existing storage solution as that is
>>>working
>>> out very well for them, but even if we did put them onto the Luster
>>>FS,  I
>>> think they would be fine.  based on that, it didn't seem so crazy to
>>> consider block access in this method.   that said,  I think we would
>>>be one
>>> of the first in M&E to do so,  pioneers if you will...
>>>
>>>
>>> diversify - we will end up in the same boat for the same reasons.
>>>
>>>
>>> thanks Charles,
>>> greg
>>>
>>>
>>>
>>>
>>>
>>>
>>> On Thu, Jan 17, 2013 at 2:20 PM, Hammitt, Charles Allen <
>>> chammitt at email.unc.edu> wrote:
>>>
>>>>  ** **
>>>>
>>>> Somewhat surprised that no one has responded yet; although it?s likely
>>>> that the responses would be rather subjective?including mine, of
>>>>course!
>>>> ****
>>>>
>>>> ** **
>>>>
>>>> Generally I would say that it would be interesting to know more about
>>>> your datasets and intended workload; however, you mention this is to
>>>>be
>>>> used as your day-to-day main business storage?so I imagine those
>>>> characteristics would greatly vary? mine certainly do; that much is
>>>>for
>>>> sure!****
>>>>
>>>> ** **
>>>>
>>>> I don?t really think uptime would be as much an issue here; there are
>>>> lots of redundancies, recovery mechanisms, and plenty of stable
>>>>branches to
>>>> choose from?the question becomes what are the feature-set needs,
>>>> performance usability for different file types and workloads, and
>>>>general
>>>> comfort level with greater complexity and somewhat less resources.
>>>>That
>>>> said, I?d personally be a bit wary of using it as a general
>>>>filesystem for
>>>> *all* your needs.  ****
>>>>
>>>> ** **
>>>>
>>>> ** **
>>>>
>>>> I do find it interesting that your short list is a wide range mix of
>>>> storage and filesystem types; traditional NAS, scale-out NAS, and
>>>>then some
>>>> block storage with a parallel filesytem in Lustre.  Why no GPFS on
>>>>the list
>>>> for comparison?****
>>>>
>>>> ** **
>>>>
>>>> I currently manage, or have used in the past *[bluearc]*, all the
>>>> storage / filesystems and more from your list.  The reason being is
>>>>that
>>>> different storage and filesystems components have some things they
>>>>are good
>>>> at? while other things they might not be as good at doing.  So I
>>>>diversify
>>>> by putting different storage/filesystem component pieces in the areas
>>>>where
>>>> they excel at best?****
>>>>
>>>> ** **
>>>>
>>>> ** **
>>>>
>>>> ** **
>>>>
>>>> Regards,****
>>>>
>>>> ** **
>>>>
>>>> Charles****
>>>>
>>>> ** **
>>>>
>>>> ** **
>>>>
>>>> ** **
>>>>
>>>> *From:* lustre-discuss-bounces at lists.lustre.org [mailto:
>>>> lustre-discuss-bounces at lists.lustre.org] *On Behalf Of *greg whynott
>>>> *Sent:* Thursday, January 17, 2013 12:18 PM
>>>> *To:* lustre-discuss at lists.lustre.org
>>>>
>>>> *Subject:* [Lustre-discuss] is Luster ready for prime time?****
>>>>
>>>>  ** **
>>>>
>>>> Hello,
>>>>
>>>>
>>>> just signed up today, please forgive me if this question has been
>>>> covered recently.  - in a bit of a rush to get an answer on this as
>>>>we need
>>>> to make a decision soon,  the idea of using luster was thrown into
>>>>the mix
>>>> very late in the decision making process.
>>>>
>>>> ****
>>>>
>>>>  We are looking to procure a new storage solution which will
>>>> predominately be used for HPC output but will also be used as our main
>>>> business centric storage for day to day use.  Meaning the file system
>>>>needs
>>>> to be available 24/7/365.    The last time I was involved in
>>>>considering
>>>> Luster was about 6 years ago and it was at that time being considered
>>>>for
>>>> scratch space for HPC usage only. ****
>>>>
>>>> Our VMs and databases would remain on non-luster storage as we already
>>>> have that in place and it works well.    The luster file system
>>>>potentially
>>>> would have everything else.  Projects we work on typically take up to
>>>>2
>>>> years to complete and during that time we would want all assets to
>>>>remain
>>>> on the file system.****
>>>>
>>>> Some of the vendors on our short list include HDS(Blue Arc), Isilon
>>>>and
>>>> NetApp.    Last week we started bouncing the idea of using Luster
>>>>around.
>>>> I'd love to use it if it is considered stable enough to do so.
>>>>
>>>> your thoughts and/or comments would be greatly appreciated.  thanks
>>>>for
>>>> your time.
>>>>
>>>> greg
>>>>
>>>>
>>>> ****
>>>>
>>>
>>>
>>> _______________________________________________
>>> Lustre-discuss mailing list
>>> Lustre-discuss at lists.lustre.org
>>> http://lists.lustre.org/mailman/listinfo/lustre-discuss
>>>
>>>
>>
>-------------- next part --------------
>An HTML attachment was scrubbed...
>URL: 
>http://lists.lustre.org/pipermail/lustre-discuss/attachments/20130121/d311
>779c/attachment-0001.html
>
>------------------------------
>
>_______________________________________________
>Lustre-discuss mailing list
>Lustre-discuss at lists.lustre.org
>http://lists.lustre.org/mailman/listinfo/lustre-discuss
>
>
>End of Lustre-discuss Digest, Vol 84, Issue 12
>**********************************************




More information about the lustre-discuss mailing list