Advatronix Cirrus 1200: a Storage Server Under Your Desk
by Johan De Gelas on June 6, 2014 5:00 AM ESTThe Specs
Below you can find the detailed specs.
Advatronix Cirrus 1200 (version 2013) | ||
CPU & RAM |
Intel Xeon E3-1265LV2 4C/8T (2.4 GHz, 8MB L3) Up to 32GB of ECC DDR3 UDIMMs dual channel, 1600 MHz |
|
Motherboard | Supermicro X9SCL | |
Storage | Drive Bays |
12 x 3.5" hot swappable (hard drive cage) Populated w 8x Seagate NAS HDD ST4000VN000 4TB—RAID-10 All HD write back caches disabled 6x 2.5'' hot swappable (SSD front drive cage) Populated w 2x Intel SSD710 200GB—RAID-1 |
Controller | Adaptec ASR71605Q with "MaxCache" and BBU Enabled | |
Cooling | Front | 80mm fan |
Rear | 2x 120mm fan | |
Top | none | |
Left Side | 80mm fan | |
Bottom | none | |
I/O Ports |
4x USB 2.0 front 2x USB 2.0 rear 2x RJ-45 Ethernet rear PS/2 mouse and Keyboard RJ-45 IPMI 2.0 Ethernet VGA D-sub Serial Com Optional : 1x RJ45 10G Ethernet |
|
Power Supply |
One 400W 80 Plus Gold PSU (not in our review unit) or Dual Redundant Athena Power 500W AP-RRMUD6508 (review unit) |
|
Case Dimensions |
Height 14" 13/16" (376mm) Width 12" 1/2" (317,5mm) Depth 12" 5.5/16" (313mm) Weight—54 lbs (24.5 kg) |
|
Prominent Features |
Cube design Two large 3.5" disk enclosure with hot swappable drives and one Pricing includes 12 SATA drives |
|
Price | starting at $4449 (with CentOS and 4GB of RAM) |
Advatronix clearly targets people with demanding storage requirements: even the low-end configuration comes with ten 2TB SATA drives (RAID-5 + one hotspare) for your data, and two 250GB SSDs in RAID-1 for your boot disks. To keep the starting price low, the server only comes with 4GB RAM, which is a bad call in our opinion. Even if you use the Advatronix as a massive capacity NAS, the extra RAM is very helpful as the OS can use the RAM as file system cache. For $150, you can get 16GB, so it's not a big deal, but it would have been better to start with two 8GB DIMMs.
39 Comments
View All Comments
thomas-hrb - Friday, June 6, 2014 - link
If you looking at storage servers under the desk why not consider something like the DELL VRTX. that at least have a significant advantage in the scalability department. You can start small and re-dimension to many different use cases as you growJohanAnandtech - Friday, June 6, 2014 - link
Good suggestion, although the DELL VRTX is a bit higher in the (pricing) food chain than the servers I described in this article.DanNeely - Friday, June 6, 2014 - link
With room for 4 blades in the enclosure the VRTX is also significantly higher in terms of overall capability. Were you unable to find a server from someone else that was a close match in specifications to the Cirrus 1200? Even if it cost significantly more, I think at least one of comparison systems should've been picked for equivalent capability instead of equivalent pricing.jjeff1 - Friday, June 6, 2014 - link
I'm not sure who would want this server. If you have a large SQL database, you definitly need more memory and better reliability. Same thing if you have a large amount of business data.Dell, HP or IBM could all provide a better box with much better support options. This HP server supports 18 disk slots, 2 12 core CPUs, and 768GB memory.
http://www8.hp.com/us/en/products/proliant-servers...
It'll cost more, no doubt. But if you have a business that's generating TBs of data, you can afford it.
Jeff7181 - Sunday, June 8, 2014 - link
If you have a large SQL database, or any SQL database, you wouldn't run it on this box. This is a storage server, not a compute server.Gonemad - Friday, June 6, 2014 - link
I've seen U server racks on wheels, with a dark glass and keys locking it, but that was just an empty "wardrobe" where you would put your servers. It was small enough to be pushed around, but with enough real estate to hide a keyboard and monitor in there, like a hypervisor KVM solution. On the plus side, if you ever decided to upgrade, just plop your gear on a real rack unit. It felt less cumbersome than that huge metal box you showed there.Then again, a server that conforms to a rack shape is needed.
Kevin G - Friday, June 6, 2014 - link
Actually I have such a Gator case. It is sold as a portable case for AV hardware but conforms to standard 19" rack mount widths and hole mounts. There is one main gotcha with my unit: it does't provide as much depth as a full rack. I have to use shorter server cases and they tend to be a bit taller. It works out as the cooling systems of taller rack cases tend to be quieter and an advantage when bring them to other locations An more of a personal preference thing but I don't use sliding rails in a portable case as I don't see that as wise for a unit that's going to be frequently moved around and traveling.martixy - Friday, June 6, 2014 - link
Someone explain something to me please.So this is specifically low-power - 500W on spec. Let's say then that it's a non-low-power(e.g. twice - 1kW). I'm gonna assume we're threading on CRAC territory at that point. So why exactly? Why would a high powered gaming rig be able to easily handle that load, even under air cooling, but a server with the same power factor require special cooling equipment with fancy acronyms like CRAC?
alaricljs - Friday, June 6, 2014 - link
A gaming rig isn't going to be pushing that much wattage 24x7. A server is considered a constant load and proper AC calculations even go so far as to consider # of people expected in a room consistently, so a high wattage computer is definitely part of the equation.DanNeely - Friday, June 6, 2014 - link
I suspect it's mostly marketing BS. One box even a high power one that's at a constant 100% load doesn't need special cooling. A CRAC is needed when you've got a data center packed full of servers because they collectively put out enough heat to overwhelm general purpose AC units. (With the rise of virtualization many older data centers capacity has become a thermal limit instead of being limited by the number of racks there's room for.)At the margin they may be saying it was designed with enough cooling to keep temps reasonable in air on the warm side of room temperature instead of only when it's being blasted with chilled air. OTOH a number of companies that have experimented with running their data centers 10 or 20F hotter than traditional have found the cost savings from cooling didn't have any major impact on longevity so...