Supermicro is a true innovator when it comes to form factors. Two years ago Supermicro invented the Twin² a 2U chassis with four servers. The 4 server nodes were fed by two shared redundant 1200W PSUs. When we heard about the Twin³, we became very curious.

The Twin³ contains 4 hotpluggable units just like the Twin². However, each unit contains two separate motherboards, and as a result there are 8 servers inside this 2U node.  The Twin³ is not aimed at the same market as the Twin². The Twin² offers high processing power in a dense, high available chassis.

The Twin³ , also known under the less sexy name Superserver 2015TA-HTRF, is based on the dual-core Intel Atom D525 1.8 GHz. Each node also supports 4 GB of non-ECC DDR3-800 SO-DIMMs, dual Gigabit Ethernet LAN, 2 USB ports, IPMI 2.0 with virtual media over LAN support, and 3x 2.5” SATA hot-swap HDDs. With 4 GB of RAM and a dual-core Atom, we can hardly speak of the Twin³ as a speed daemon.

It is not completely clear to me which market is targeted. According to Supermicro, there are server markets where very low power in idle is important and where little processing power is required. I would personally solve that problem by consolidating all those low intensive workloads on a decently powerful virtualized host with tons of RAM so that you can keep the total amount of servers low. But we have to give this server the benefit of the doubt. 

The system also includes (1+1) redundant 720W Gold Level power supplies with PMBus, and four high quality fans. In reality it will be hard to push this system into consuming more than 400W.

HTX versus PCIe More GPU Processing Power
Comments Locked

17 Comments

View All Comments

  • rahvin - Monday, March 14, 2011 - link

    And why exactly would you even want Tegra (regardless of version) in a server? It brings nothing to a server, Atom has a use, low power, x86 compatibility and is ideal for low use, high IO workloads. What niche would Tegra serve in a rackmount server?
  • zhill - Monday, March 14, 2011 - link

    Good discussion of low-power servers and where they make sense and don't from James Hamilton at Amazon Web Services (this is his personal blog, not official Amazon):
    http://perspectives.mvdirona.com/2010/05/18/WhenVe...

    "Where very low-power, low-cost servers win is:
    1. Very cold storage workloads....The core challenge with cold storage apps is that overall system cost is dominated by disk but the disk needs to be attached to a server. We have to amortize the cost of the server over the attached disk storage. The more disk we attach to a single server, the lower the cost. But, the more disk we attach to a single server, the larger the failure zone. Nobody wants to have to move 64 to 128 TB every time a server fails. The tension is more disk to server ratio drives down costs but explodes the negative impact of server failures. So, if we have a choice of more disks to a given server or, instead, to use a smaller, cheaper server, the conclusion is clear. Smaller wins. This is a wonderful example of where low-power servers are a win.
    2. Workloads with good scaling characteristics and non-significant local resource requirements. Web workloads that just accept connections and dispatch can run well on these processors. However, we still need to consider the “and non-significant local resource” clause. If the workload scales perfectly but each interaction needs access to very large memories for example, it may be poor choice for Wimpy nodes. If the workload scales with CPU and local resources are small, Wimpy nodes are a win."
  • rahvin - Monday, March 14, 2011 - link

    The other key thing is there are a few circumstances that make virtualization a very bad choice. Virtual severs handles heavy IO very poorly with significant latency even if the individual workloads are terribly small. This is an area where the Atom servers are big sellers. Negligible power load, small form factor (1U, 18" deep) and a full hardware even if it's crappy hardware. One of the models comes with two mini-itx atom boards in the same 1U and they each have their own hard drive.

    This is a niche market, but based on Supermicro producing several models I bet they are selling quite a few. I'd be curious to find out if Supermicro would reveal how many of the Atom servers they sell. I know there is quite a bit of discussion around about these things being the perfect custom firewall (one of those heavy IO loads) on a corporate network.
  • mino - Monday, March 14, 2011 - link

    "However, it seems that the HTX slot is at the end of its lifetime. The upcoming Xeons seem to come with a PCI-express 3.0 controller integrated, so they should be able to offer a low latency interface of up to 12.8 GB/s, or twice as much."

    AFAIR Hyper Transport has, thanks to the communication protocol itself, about 1/2 the latency of PCI. Regardless the physical layer.
  • JMC2000 - Tuesday, March 15, 2011 - link

    That's what I didn't understand. HT3.0 has lower latency and higher bandwidth than PCI-E 3.0, so how can it be near the end of its life? HTX3.0 has been in use ever since the launch of C32/G34.
  • jcandle - Wednesday, March 16, 2011 - link

    Actually the article is quite correct. There are a limited number of HTX expansion cards produced and since most of these cards are designed for specialized tasks. There's a considerably small market for these cards to begin with. Its easier to kill off HTX in favor of a PCI-E where the same card can be used in both Intel and AMD platforms. With PCI-E 3.0, HTX is essentially dead. Now, that doesn't mean HT is dead; only HTX. Even now with Infiniband cards there are far better optimization to be made on the software side to increase performance than eeking out the very drop of latency from HTX over PCI-E.
  • jcandle - Wednesday, March 16, 2011 - link

    Anyone notice the incorrect specs. According to their datasheet the 6F+ has only two slots the HTX and 1 PCI-E. If its like the usual Supermicro board the remaining slots should be unpopulated. The 6F as it appears has only 68 PCI-E lanes not 80 indicated in the article. 5 full lane PCI-E 16x slots made for anything less than a 4U/workstation is a considerably more rare.

    H8QGL-6F/H8QGL-iF:
    3 PCI-Express 2.0 x16
    2 PCI-Express 2.0 x8 (using x16 slot)
    1 PCI-Express 2.0 x4 (using x16 slot)

    H8QGL-6F+/H8QGL-iF+:
    1 HyperTransport slot
    1 PCI-Express 2.0 x16

Log in

Don't have an account? Sign up now