[Lustre-discuss] Lustre IOkit newbie: sgpdd-survey
Ms. Megan Larko
dobsonunit at gmail.com
Tue Jul 29 12:38:18 PDT 2008
Hello,
I'm back to working on benchmarking the hardware prior to lustre fs
installation.
I have used the LSI 8888ELP WebBIOS utility to destroy the previous
RAID configuration on one of our 16-bay JBOD units. A new set of
RAID arrays were set up for benchmarking purposes. One RAID6 with
parity disks, one RAID5 with parity disk and one single spindle
passed directly to the OSS. The boot of CentOS 5
(2.6.18-53.1.13.el5_lustre.1.6.4.3smp) detects these new block devices
as /dev/sdh, /dev/sdi and /dev/sdj respectively. The "sg_map" command
detects the raw units. The lustre-IOkit is still unable to find
them. Themodule "sg" is indeed loaded as is the megaraid_sas for the
LSI 8888ELP card. Can lustre-IOkit actually benchmark anything on an
LSI 8888ELP card? Should I use another tool like IOzone?
Thanks,
megan
>From dmesg:
sd 2:2:1:0: Attached scsi generic sg12 type 0
Vendor: LSI Model: MegaRAID 8888ELP Rev: 1.20
Type: Direct-Access ANSI SCSI revision: 05
sdg : very big device. try to use READ CAPACITY(16).
SCSI device sdg: 11707023360 512-byte hdwr sectors (5993996 MB)
sdg: Write Protect is off
sdg: Mode Sense: 1f 00 00 08
SCSI device sdg: drive cache: write back, no read (daft)
sdg: unknown partition table
sd 2:2:2:0: Attached scsi disk sdg
sd 2:2:2:0: Attached scsi generic sg13 type 0
Vendor: LSI Model: MegaRAID 8888ELP Rev: 1.20
Type: Direct-Access ANSI SCSI revision: 05
sdh : very big device. try to use READ CAPACITY(16).
SCSI device sdh: 11707023360 512-byte hdwr sectors (5993996 MB)
sdh: Write Protect is off
sdh: Mode Sense: 1f 00 00 08
SCSI device sdh: drive cache: write back, no read (daft)
sdh: unknown partition table
sd 2:2:3:0: Attached scsi disk sdh
sd 2:2:3:0: Attached scsi generic sg14 type 0
Vendor: LSI Model: MegaRAID 8888ELP Rev: 1.20
Type: Direct-Access ANSI SCSI revision: 05
SCSI device sdi: 1951170560 512-byte hdwr sectors (998999 MB)
sdi: Write Protect is off
sdi: Mode Sense: 1f 00 00 08
SCSI device sdi: drive cache: write back, no read (daft)
SCSI device sdi: 1951170560 512-byte hdwr sectors (998999 MB)
sdi: Write Protect is off
sdi: Mode Sense: 1f 00 00 08
SCSI device sdi: drive cache: write back, no read (daft)
sdi: unknown partition table
sd 2:2:4:0: Attached scsi disk sdi
sd 2:2:4:0: Attached scsi generic sg15 type 0
[root at oss4 ~]# sg_map
/dev/sg0 /dev/sda
/dev/sg1 /dev/scd0
/dev/sg2
/dev/sg3
/dev/sg4
/dev/sg5 /dev/sdb
/dev/sg6 /dev/sdc
/dev/sg7 /dev/sdd
/dev/sg8
/dev/sg9
/dev/sg10
/dev/sg11 /dev/sde
/dev/sg12 /dev/sdf
/dev/sg13 /dev/sdg
/dev/sg14 /dev/sdh
/dev/sg15 /dev/sdi
/dev/sg16 /dev/sdj
/dev/sg17 /dev/sdk
Lustre-IOkit sgpdd-survey:
scsidevs=/dev/sg16
Result of sgpdd-survey:
[root at oss4 log]# sgpdd-survey
Can't find SG device for /dev/sg16, testing for partition
Can't find SG device /dev/sg1.
Do you have the sg module configured for your kernel?
[root at oss4 log]# lsmod | grep sg
sg 70056 0
scsi_mod 187192 11
ib_iser,libiscsi,scsi_transport_iscsi,ib_srp,sr_mod,libata,megaraid_sas,sg,3w_9xxx,usb_storage,sd_mod
This is still using the "-16" option to each of the two sg_readcap calls.
sg_readcap -b -16
********************************************************************
On Thu, Jul 24, 2008 at 3:45 PM, Andreas Dilger <adilger at sun.com> wrote:
> On Jul 24, 2008 14:52 -0400, Ms. Megan Larko wrote:
>> Okay. So I have a JBOD with 16 1Tb Hitachi Ultrastar sATA drives in
>> it connected to the OSS via an LSI 8888ELP controller card (with
>> Battery Back-up). The JBOD is passed to the OSS as raw space. My
>> partitions cannot exceed 8Tb, but splitting into two 7Tb will hurt
>> performance....
>
> The ideal layout would be to have two RAID-5 arrays with 8 data + 1 parity
> disks using 64kB or 128kB RAID chunk size, but you are short two disks...
>
> You may also consider using RAID-6 with 6 data + 2 parity disks. Having
> the RAID stripe width be 15+1 is probably quite bad because a single
> 1MB RPC will always need to recalculate the parity for that IO.
>
>> What do I do for Lustre set-up? I thought the fewer partitions the
>> better because one does has less "overhead" space. Do I put them out
>> as 16 single Tb partitions????? That seems like extra work for a
>> file system to track.
>
> Please see section 10.1 in the Lustre manual for more tips:
> http://manual.lustre.org/manual/LustreManual16_HTML/RAID.html
>
> Cheers, Andreas
> --
> Andreas Dilger
> Sr. Staff Engineer, Lustre Group
> Sun Microsystems of Canada, Inc.
>
>
More information about the lustre-discuss
mailing list