Final Words

The GIGABYTE GB-BXi7H-5500 provided us with the opportunity to take a look at what Broadwell-U can deliver. The migration from 22nm to 14nm has allowed for higher base clocks while maintaining the same power envelop. The performance delta over the Haswell-U-based GB-BXi7-4500 (particularly, on the graphics side) is notable. That said, while migrating from Sandy Bridge or Ivy Bridge is a no-brainer, there is not enough on offer to recommend migrating from a Haswell-based mini-PC.

From the perspective of the BRIX itself, the GB-BXi7H-5500 has two major downsides, and both of them are related to the networking subsystem. Considering that the i7 SKU is the 'premium' SKU in this lineup, use of a 1x1 802.11ac (Intel AC3160) card is disappointing. It would have been nice to have the AC7260 or AC7265. The Broadcom BCM4352 is even more preferable. In addition, GIGABYTE should have opted for an Intel LAN chipset instead of a Realtek one for this high-end configuration.

One of the interesting differentiating features is the availability of NFC. Along with the unit, GIGABYTE also supplied a NFC tag. The accompanying driver DVD has a NFC app (Windows program) which allows creation of profiles to link with the tag. These profiles could be associated with web addresses or files / programs in the system. Tapping the tag against the unit's lid activates the profile. Right now, this is a basic application. GIGABYTE indicated that they are working on bringing more functionality to the app and making it easier to use.

The pricing of the Core i7-5500U makes it attractive for high-end ultrabooks, but the BXi7H-5500 manages to get it to the consumer in a UCFF PC at a reasonable price ($509, when Intel apparently sells the CPU alone for $393). All in all, GIGABYTE has managed to deliver the right balance of price, power, size and thermal / acoustic design in the GB-BXi7H-5500. It remains to be seen what Intel is able to offer with the NUC lineup and how the NUCs will complement and/or compete against the Broadwell BRIX units.

Power Consumption and Thermal Performance
Comments Locked

53 Comments

View All Comments

  • gonchuki - Thursday, January 29, 2015 - link

    Please define "decimates". In non-OpenCL and non-GPU bound tests, it's a 5-7% win at most, which can be easily explained by the 33% higher base clock of the CPU cores, plus the die shrink that allows for better thermals (more headroom for higher bins of turbo boost).

    All of the test results point to Broadwell having the exact same IPC as Haswell in all situations. If anything improved it can only be because of the new stepping that might have fixed some errata.
  • nathanddrews - Thursday, January 29, 2015 - link

    Agreed, the word "decimates" is a bit extreme - but I consider anything in the 10-20% range to be significantly better.
  • Refuge - Thursday, January 29, 2015 - link

    These days I agree, long gone are the days of Sandybridge... Tis a shame, they were fun.
  • Laststop311 - Friday, January 30, 2015 - link

    I'm glad some 1 else noticed this. In the benchmarks that strictly use only the cpu, broadwells haswell equivalent is barely and i mean barely any faster. Sure the gpu is a pretty decent improvement but who cares about intels integrated gpu's? Anyone that relies heavily on an integrated gpu is going to get an apu from amd. The only reason the gpu is so much better is its such a poor performing part to begin with, it's a lot easier to improve lower performing things than things that are already highly optimized like the cpu.

    This is bad news for people using desktops with discrete gpu's and were hoping broadwell would be a decent boost. In those situations the iGPU means nothing so big deal it got better. This also means broadwell-e is going to rly suck and be basically identical to haswell-e almost no reason to even bother designing broadwell-e chips since they dont even use iGPU there is no performance increase at all to talk about in those.

    The silver lining though is we get to save money another year. With intel having no pressure on them we get to save our money till there is a real performance boost. Basically anyone with an i7-920 or higher doesn't have to spend money on a pc upgrade till maybe skylake/skylake-e MAYBE, intel has put out underwhelming tocks lately as well. My x58 i7-980x system still has no cpu bottleneck. This allowed me to buy a 55" LG OLED tv as normally i was buying a new pc every 2-3 years before the core i7 series started then all the sudden performance upgrades became pathetic, my new pc fund built up and i found the oled tv for 3000 and figured why not i can easily go another couple years with the same pc. So thanks intel for making no progress i got a new oled tv.
  • BrokenCrayons - Friday, January 30, 2015 - link

    I care about iGPU benchmarks and the computer I use for gaming has an Intel HD3000 and probably will do so for at least another year or more before even thinking about an upgrade. Having dedicated graphics in my laptop seems pointless when I can just wait 5-7 years or so to play a game after it's fully patched and usually avaiable with all of it's DLC for very little cost plus runs well on something that doesn't need a higher end graphics processor. So yes, for serious gaming, iGPUs are fine if you manage expectations and play things your computer can easily handle.
  • purerice - Saturday, January 31, 2015 - link

    BrokenCrayons, agreed 100%!! I recently upgraded from Merom to Ivy Bridge myself.

    There are tons of games now selling for $5-$10 that wouldn't run on Merom when the games cost $40-$60. In addition to being patched and DLC'd, guides and walkthroughs exist to get through any of the "less awesome" parts. More money saved for real life and less frustration to interrupt gaming. Patience pays indeed.

    Seeing the ~20% boost over 4500U in Ice Storm and Cinebench Open GL was actually exciting, even if it represents performance below 90% of other Anandtech users' current levels.
  • DrMrLordX - Tuesday, February 3, 2015 - link

    Decimates means to destroy something by 10% of its whole. All things considered, I'd rather be decimated than . . . you know, devastated, or annihilated.
  • DanNeely - Thursday, January 29, 2015 - link

    While I agree that replacing 1280x1024 is past due; I disagree with picking 1280x720. Back when it was picked 1280x1024 was the most common resolution on low end monitors. Today the default low end resolution is 1366x768 (26.65% on steam); it's also the second most commonly used one (after 1080p).
  • Oxford Guy - Thursday, January 29, 2015 - link

    Agreed. It's pretty silly to "replace" a higher resolution with a lower one.
  • frozentundra123456 - Saturday, January 31, 2015 - link

    I would disagree. It is quite reasonable, because many laptops use 768p, as well as cheap TVs.

Log in

Don't have an account? Sign up now