The Chassis is a Dell Blade Chassis m1000e with an 8-port Brocade fiber switch and an 8-port Cisco 3130 copper switch.
I am sharing the Chassis with about 10 Dell M610 half-height, medium load servers. I will have the only machines with access to the fibre channel cards. The rest use the copper interconnects. I can get a VLAN on the copper as needed, and most likely can ask for higher priority traffic routing on the copper.
SAN backend with EMC Clariion, 2 trays of disks, 15 15K 300 GB spindles each.
Must be able to cluster the three machines for high availability. I would prefer an active/active/active (scalable) cluster config, but would consider failover if the right argument can be made.
I have the choice of using any HA/Clustering operating system that will install on these M905’s, so long as
a support contract can be bought for the OS (the manager requires this)
all devices are supported with minimal fuss by installer (with minimal repackaging, as needed)
it can export NFS shares and Samba (I would do straight NFS, but there are non-Windows 7 desktops still in use)
So far, I have
one person recommending to use RedHat 5.5 with GFS2, clvmd and a lump of SAN presented to the cluster for use.
one person recommending Solaris 10 10/9 with ZFS and forcing one LUN per disk/JBOD presentation from the SAN.
A third, more “introspective” party told me to contract it out and have someone to blame when it does not work (not an option).
I need to support 1000+ users (> 250 concurrent on average) of NFS and Samba (it’s literally just an HA/Cluster for file sharing, no web servers or applications/DBs/etc.)
My SAN admin is a great guy and will accommodate my requests (even if it’s more work for him to do JBOD presentations).
What would you do with this setup to satisfy up to 1000+ users?
(BTW, a valid answer could include “hire a consultant to analyze the situation and give a report on how to proceed”)
I think I understood the first sentence. I’m not exactly sure what you’re doing with this, but if I can’t download ANSYS 13.0 when it comes out, I’m blaming you.
Depending on the amount of storage you need and projected growth you should design you LUN or LUNs also if your storing video or a certain type of file you can tune stripe breadth for max performance
Jeff95TA, these are not external facing servers, no worries. But I can yell over the wall at the guy responsible!
LZ, I don’t really know why I was given these three particular machines. I thought it was overkill too. I’d be better off using them for HPC solver testing or a DB middleware server than file serving, but oh well. We literally had them laying around and they’re under warranty still, so what the heck.
And the idea of not letting the Clariion handle the striping was a baffling approach…so much so that I figured a post to the new neighborhood was in order to see if any SAN admins out there had their spidey-sense start tingling…apparently it did!
The data is binary and dense (.zip, .tar.gz, .rpm, sun flash archives, satellite kickstarts, autoyast2 images, .iso, etc.) individual file sizes range from 1KB to 400GB (a rather wide variety) with a large deviation in sizes.
Everything I am reading tells me that using spreading the disks over 5-6 LUNS ranging from 500 GB to 800 GB in size would probably be a good compromise and leave a few spares in the array for expansion.
The actual underlying data files will only grow about 10% a year, so by the time I need more disks these Clariions and the Blades will be out of warranty and I’ll get to start over in a few years with a new system.
I’m hoping solid state will come down in price and I can go with a tray of solid state drives.
I think this wins as one of the most technical questions asked on here.
Depending on amount of data you need stored usually you build X number of LUNs then stripe across the LUNs and let the SAN handle all that…also leaving room for hot spares.
Since you’re running a variety of file sizes and types there isn’t that much tuning you can do.
Thanks, I appreciate the help and input. I’m going to grab a few disks and config them in various ways, profiling the copying of variously sized files from many scripted, virtual clients spread over a pretty large vSphere. Should give me somewhere to start.
Would solid-state flash-caches help with the I/O bottleneck I will ultimately run into at the Clariion?
I can’t believe what a difference this setup makes. Using the flash-cache and RAID 5 LUNS on Linux w/xfs/gfs/gfs2 using either Linux or EMC multipathing can’t even touch this.
The hardcore Solaris geeks I talk to agree about the Oracle thing but they also agree that the ability to dump a supported Solaris on to certified Dell x86 PC buckets is probably the one thing that will prevent Solaris from going the way of the dinosaur, like xfs/SGI and hpfs/HPUX and let’s not forget DEC.
Next step is to bond the 4 copper ports on the M905 and have 4 1Gb/s front ends funnelling down into the system and test reads. Once the 128 GB RAM for ARC warms up, I expect even better things.