[Lustre-discuss] is Luster ready for prime time?

greg whynott greg.whynott at gmail.com
Mon Jan 21 08:24:32 PST 2013


Thanks very much Indivar,  informative read.    it is good to see others in
our sector are using the technology and you have some good points.

have a great day,
greg



On Sat, Jan 19, 2013 at 6:52 AM, Indivar Nair <indivar.nair at techterra.in>wrote:

>  Hi Greg,
>
> One of our customers had a similar requirement and we deployed Lustre
> 2.0.0.1 for them. This was in July 2011. Though there were a lots of
> problems initially, all of them were sorted out over time. They are quite
> happy with it now.
>
> *Environment:*
> Its a 150 Artist studio with around 60 Render nodes. The studio mainly
> uses Mocha, After Effects, Silhouette, Synth Eye, Maya, and Nuke among
> others. They mainly work on 3D Effects and Stereoscopy Conversions.
> Around 45% of Artists and Render Nodes are on Linux and use native Lustre
> Client. All others access it through Samba.
>
> *Lustre Setup:*
> It consists of 2 x Dell R610 as MDS Nodes, and 4 x Dell R710 as OSS Nodes.
> 2 x Dell MD3200 with 12x1TB SAS Nearline Disks are used for storage. Each
> Dell MD3200s are shared among 2 OSS nodes for H/A.
>
> Since the original plan (which didn't happen) was to move to a 100% Linux
> environment, we didn't allocate separate Samba Gateways and use the OSS
> nodes with CTDB for it. Thankfully, we haven't had any issues with that yet.
>
> *Performance:*
> We get a good THROUGHPUT of 800 - 1000MB/s with Lustre Caching. The disks
> it self provide much lesser speeds. But that is fine, as caching is in
> effect most of the time.
>
> *Challenge:*
> The challenge for us was to tune the storage for small files 10 - 50MB
> totalling to 10s of GBs. An average shot would consist of 2000 - 4000  .dpx
> images. Some Scenes / Shots also had millions of <1MB Maya Cache files.
> This did tax the storage, especially the MDS. Fixed it to an extent by
> adding more RAM to MDS.
>
> *Suggestions:*
>
> 1. Get the real number of small files (I mean <1MB ones) created / used by
> all software. These are the ones that could give you the most trouble. Do
> not assume anything.
>
> 2. Get the file - sizes, numbers and access patterns absolutely correct.
> This is the key.
>     Its easier to design and tune Lustre for large files and I/O.
>
> 3. Network tuning is as important and storage tuning. Tune Switches, each
> Workstation, Render Nodes, Samba / NFS Gateways, OSS Nodes, MDS Nodes,
> everything.
>
> 4. Similarly do not undermine Samba / NFS Gateway. Size and tune them
> correctly too.
>
> 5. Use High Speed Switching like QDR Infiniband or 40GigE, especially for
> backend connectivity between Samba/NFS Gateway and Lustre MDS/OSS Nodes.
>
> 6. As far as possible, have fixed directory pattern for all projects.
> Separate working files (Maya, Nuke, etc.) from the data, i.e. frames /
> images, videos, etc. at the top directory level it self. This will help you
> tune / manage the storage better. Different directory tree for different
> file sizes or file access types.
>
> If designed and tuned right, I think Lustre is best storage currently
> available for your kind of work.
>
> Hope this helps.
>
> Regards,
>
>
> Indivar Nair
>
>
> On Fri, Jan 18, 2013 at 1:51 AM, greg whynott <greg.whynott at gmail.com>wrote:
>
>> Hi Charles,
>>
>>   I received a few off list challenging email messages along with a few
>> fishing ones,  but its all good.   its interesting how a post asking a
>> question can make someone appear angry.  8)
>>
>> Our IO profiles from the different segments of our business do vary
>> greatly.   The HPC is more or less the typical load you would expect to
>> see,  depending on which software is in use for the for the job being ran.
>>       We have hundreds of artists and administrative staff who use the file
>> system in a variety of ways.   Some examples would include but not limited
>> to:  saving out multiple revisions of photoshop documents (typically in the
>> hundreds of megs to +1gig range),   video editing (stereoscopic 2k and 4k
>> images(again from 10's 100's to gigs in size) including uncompressed
>> video,  excel, word and similar files,  thousands of project files (from
>> software such as Maya,  Nuke and similar)  these also vary largely in size,
>> from 1 to thousands of megs in size.
>>
>> The intention is keep our data bases and VM requirements on the existing
>> file system which is comprised of about 100 10k SAS drives,  it works well.
>>
>> We did consider GPFS but that consideration went out the door once I
>> started talking to them and hammering in some numbers into their online
>> calculator.  Things got a bit crazy quickly.   They have different pricing
>> for the different types and speeds of Intel CPUs.  I got the feeling they
>> were trying to squeeze every penny out of customers they could.  felt very
>> Brocade-ish and left a bad taste with us.   wouldn't of been much of a
>> problem as some other shops I've worked at,  but here we do have a finite
>> budget to work within.
>>
>> The NAS vendors could all be considered scale out I suspect.   All 3 can
>> scale out the storage and front end.  NA C-mode can have up to 24 heads,
>> Blue Arc goes up to 4 or 8 depending on the class,  Isilon can go up to 24
>> nodes or more as well if memory serves me correctly,  and they all have a
>> single name space solution in place.   They each have their limits,   but
>> for our use case they are really subjective.   We will not hit the limits
>> of their scalability before we are considering a fork lift refresh.  In our
>> view,  for what they offer it is perty much a wash for them - any would
>> meet our needs.  NetApp still has a silly agg/vol size limit,  at least it
>> is up to 90TB now (from 9 in the past(formatted fs use))..  in April it is
>> suppose to go much higher.
>>
>>  The block storage idea in the mix - since all our HPC is linux,  they
>> all would become luster clients.   To provide a gateway into the luster
>> storage for none linux/luster hosts the thinking was a clustered pair of
>> linux boxes running SAMBA/NFS which were also Luster clients.    Its just
>> an idea being bounced around at this point.  The data serving requirements
>> of the non HPC parts of the business are much less.   The video editors
>> most likely would stay on our existing storage solution as that is working
>> out very well for them, but even if we did put them onto the Luster FS,  I
>> think they would be fine.  based on that, it didn't seem so crazy to
>> consider block access in this method.   that said,  I think we would be one
>> of the first in M&E to do so,  pioneers if you will...
>>
>>
>> diversify - we will end up in the same boat for the same reasons.
>>
>>
>> thanks Charles,
>> greg
>>
>>
>>
>>
>>
>>
>> On Thu, Jan 17, 2013 at 2:20 PM, Hammitt, Charles Allen <
>> chammitt at email.unc.edu> wrote:
>>
>>>  ** **
>>>
>>> Somewhat surprised that no one has responded yet; although it’s likely
>>> that the responses would be rather subjective…including mine, of course!
>>> ****
>>>
>>> ** **
>>>
>>> Generally I would say that it would be interesting to know more about
>>> your datasets and intended workload; however, you mention this is to be
>>> used as your day-to-day main business storage…so I imagine those
>>> characteristics would greatly vary… mine certainly do; that much is for
>>> sure!****
>>>
>>> ** **
>>>
>>> I don’t really think uptime would be as much an issue here; there are
>>> lots of redundancies, recovery mechanisms, and plenty of stable branches to
>>> choose from…the question becomes what are the feature-set needs,
>>> performance usability for different file types and workloads, and general
>>> comfort level with greater complexity and somewhat less resources.  That
>>> said, I’d personally be a bit wary of using it as a general filesystem for
>>> *all* your needs.  ****
>>>
>>> ** **
>>>
>>> ** **
>>>
>>> I do find it interesting that your short list is a wide range mix of
>>> storage and filesystem types; traditional NAS, scale-out NAS, and then some
>>> block storage with a parallel filesytem in Lustre.  Why no GPFS on the list
>>> for comparison?****
>>>
>>> ** **
>>>
>>> I currently manage, or have used in the past *[bluearc]*, all the
>>> storage / filesystems and more from your list.  The reason being is that
>>> different storage and filesystems components have some things they are good
>>> at… while other things they might not be as good at doing.  So I diversify
>>> by putting different storage/filesystem component pieces in the areas where
>>> they excel at best…****
>>>
>>> ** **
>>>
>>> ** **
>>>
>>> ** **
>>>
>>> Regards,****
>>>
>>> ** **
>>>
>>> Charles****
>>>
>>> ** **
>>>
>>> ** **
>>>
>>> ** **
>>>
>>> *From:* lustre-discuss-bounces at lists.lustre.org [mailto:
>>> lustre-discuss-bounces at lists.lustre.org] *On Behalf Of *greg whynott
>>> *Sent:* Thursday, January 17, 2013 12:18 PM
>>> *To:* lustre-discuss at lists.lustre.org
>>>
>>> *Subject:* [Lustre-discuss] is Luster ready for prime time?****
>>>
>>>  ** **
>>>
>>> Hello,
>>>
>>>
>>> just signed up today, please forgive me if this question has been
>>> covered recently.  - in a bit of a rush to get an answer on this as we need
>>> to make a decision soon,  the idea of using luster was thrown into the mix
>>> very late in the decision making process.
>>>
>>> ****
>>>
>>>  We are looking to procure a new storage solution which will
>>> predominately be used for HPC output but will also be used as our main
>>> business centric storage for day to day use.  Meaning the file system needs
>>> to be available 24/7/365.    The last time I was involved in considering
>>> Luster was about 6 years ago and it was at that time being considered for
>>> scratch space for HPC usage only. ****
>>>
>>> Our VMs and databases would remain on non-luster storage as we already
>>> have that in place and it works well.    The luster file system potentially
>>> would have everything else.  Projects we work on typically take up to 2
>>> years to complete and during that time we would want all assets to remain
>>> on the file system.****
>>>
>>> Some of the vendors on our short list include HDS(Blue Arc), Isilon and
>>> NetApp.    Last week we started bouncing the idea of using Luster around.
>>> I'd love to use it if it is considered stable enough to do so.
>>>
>>> your thoughts and/or comments would be greatly appreciated.  thanks for
>>> your time.
>>>
>>> greg
>>>
>>>
>>> ****
>>>
>>
>>
>> _______________________________________________
>> Lustre-discuss mailing list
>> Lustre-discuss at lists.lustre.org
>> http://lists.lustre.org/mailman/listinfo/lustre-discuss
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20130121/d311779c/attachment.htm>


More information about the lustre-discuss mailing list