Idle Power Measurement

SATA SSDs are tested with SATA link power management disabled to measure their active idle power draw, and with it enabled for the deeper idle power consumption score and the idle wake-up latency test. Our testbed, like any ordinary desktop system, cannot trigger the deepest DevSleep idle state.

Idle power management for NVMe SSDs is far more complicated than for SATA SSDs. NVMe SSDs can support several different idle power states, and through the Autonomous Power State Transition (APST) feature the operating system can set a drive's policy for when to drop down to a lower power state. There is typically a tradeoff in that lower-power states take longer to enter and wake up from, so the choice about what power states to use may differ for desktop and notebooks.

We report two idle power measurements. Active idle is representative of a typical desktop, where none of the advanced PCIe link or NVMe power saving features are enabled and the drive is immediately ready to process new commands. The idle power consumption metric is measured with PCIe Active State Power Management L1.2 state enabled and NVMe APST enabled if supported.

Active Idle Power Consumption (No LPM)Idle Power Consumption

The active idle power consumption of the Phison S10 controller has always been pretty good, though the Seagate BarraCuda draws a bit more at idle than some of the older S10 drives we've tested. When SATA link power management is enabled and the drive is put to sleep, the BarraCuda's power consumption is only cut in half and every other drive with working power management comes out ahead.

Idle Wake-Up Latency

The Seagate BarraCuda had relatively poor idle power management, so it's appropriate for it to have one of the quickest wake-up latencies, slightly faster than the two older Phison S10 drives featured here. The BarraCuda wakes up seven times faster than the fastest Silicon Motion-based drive, and 26 times faster than the Toshiba TR200 that uses Phison's DRAMless S11 controller. The BarraCuda should definitely be used with the OS configured for the most aggressive drive power management, because the performance impact is minimal.

Mixed Read/Write Performance Conclusion
Comments Locked

39 Comments

View All Comments

  • seamonkey79 - Friday, December 7, 2018 - link

    The caddy would adapt an m.2 to a full 2.5" SATA socket, so you would have an m.2 in a chassis adapting it to 2.5" & classic SATA.

    That being said, I can't see that doing anything but increasing costs, though having one primary line to manufacture m.2 and the little bit of work needed to adapt an m.2 SATA drive to 2.5", it *could* see some benefit to the manufacturer.
  • Death666Angel - Friday, December 7, 2018 - link

    Huh? M.2 SATA doesn't overheat, that's M.2 NVME/PCIe (which is a bit more difficult to adapt to 2.5" SATA, so I don't care as much). 2.5" SATA drives already have tiny PCBs. They are already bottlenecked by the protocol. With them being M.2 in a 2.5" caddy, they can serve double purpose. I just bought an M.2 drive and a caddy for my Fujitsu T904 laptop, which still has only a 2.5" slot. But I know when that laptop is gone, I'm not gonna need a 2.5" drive. That M.2 drive can be converted into all kinds of useful devices, small desktops, laptops, USB thumb drive. It'd cost them a couple dozens of cents more to manufacture, would it'd be soo much more useful. :D Eventually, more people will use M.2 SATA than 2.5". At least those who buy standalone drives.
  • CheapSushi - Saturday, December 8, 2018 - link

    Dragonstongue I don't think you really know what the OP is talking about, just how tiny the PCB seems to be on newer 2.5" SSDs, thus just being a whole lot of waste of space anyway and/or you don't realize adapters already exist and/or don't realize M.2 is just a formfactor and can be NVMe/PCIe or AHCI/SATA. There's always someone who confidently posts a "neg" at a suggestion without even knowing hardware much in the first place.
  • dgingeri - Friday, December 7, 2018 - link

    When I was a systems admin for a server software test lab, we received a set (108) of 3TB Seagate Constellation ES.2 SAS drives for a new prototype appliance that would eventually become the DXi6900 series. I was pretty excited to see the new hardware come in, so I got things set up within a day.

    The test team wasn't ready to test for another week. In that week, we had 12 of those drives go bad. By two months into the project, every single Seagate drive had gone bad. (These are their enterprise level drives, which are supposed to have better reliability.) We ended up requesting a different brand drive from NetApp (the maker of the storage portion of the DXi6900) because of these problems. We replaced them with HGST 4TB drives, and didn't have a single one fail up to the point when I left over two years later.

    In the years leading up to that, I had bought several Seagate drives, including 4 1TB drives, 2 2TB drives, and one 3TB drive, and had the drive fail within warranty in EVERY SINGLE CASE. That was specifically why I quit buying them, and the 3TB drive is the last Seagate drive I am ever going to buy intentionally.

    This just might be a decent crive, and if Seagate were to put a concerted effort into improving their reliability, they might be something I'd consider. However, as things stand, Seagate and Toshiba are on my NEVER BUY list, along side Biostar, ECS, and Gigabyte.
  • CheapSushi - Saturday, December 8, 2018 - link

    What does this have to do with SSDs at all? Seriously? This is just some overreaching rant saying that one specific type of product means that ALL their products are a problem. Especially about spinning rust several years old; even BackBlaze doesn't use those. Have you even bothered to look into reliability statistics/information on even consumer TLC drives? It's so odd to see people who claim to be in tech but are so antiquated in their knowledge.
  • gglaw - Sunday, December 9, 2018 - link

    Quite a humorous post from someone with a supposed heavy tech/admin background but so short-sighted on the big picture. He lists a bad experience from a completely unrelated product line likely not even sharing manufacturing or R&D ties in any way making him ban products from some of the largest tech companies in the world with for the most part tremendous track records. Even if it is in "principle" for how the company leaders model their QC, all the executives making these decisions at the time of the archaic hard drive problems are likely working with other companies by now (pretty good chance for one of his "new" favorite companies). Similar to the comments on some of Samsung's early SSD fiascos banning all Samsung products "for life." And of course shortly after their fiascos, they quickly became essentially the world's benchmark for performance and reliability in this same product line lol.
  • Donkey2008 - Thursday, December 13, 2018 - link

    108 enterprise-class hard drives failed in 2 months?

    [Insert Doubt meme]
  • sarahkevin - Friday, December 7, 2018 - link

    thanks for sharing I really need this for my office.
  • PeachNCream - Friday, December 7, 2018 - link

    I think I'd probably opt for a Crucial MX500 over a Seagate SSD. Seagate's reputation and my experiences with their mechanical drives make me reluctant about giving them yet another chance.
  • Fujikoma - Friday, December 21, 2018 - link

    I feel the same way about Quantum SCSI drives. Not that Seagate rates much higher...

Log in

Don't have an account? Sign up now