WiFi Testing with Ixia IoT

As previously discussed, RF testing has always been a major unknown to some extent because environmental factors make it extremely difficult to tell exactly what is going on with the system. I don’t think it really needs to be said but previous reviews and any controversy regarding the quality of RF has always lead to a ring of confusion and back and forth with no clear-cut answers, at least in the public domain. The Transformer Prime and Pixel C reception issues have all been cases where I’ve seen a lot of confusion over whether a problem really existed in the hardware, software, or with the end user.

Most people really don’t have any understanding of how wireless transmission works, probably because it’s not really something you can see. As far as I know, no one is capable of seeing radio waves, even at high frequencies like 60 GHz. Of course, the problem is that for quite some time our testing was also not really ideal for seeing the quality of an RF implementation. While iPerf does provide some useful data, free space testing means that we’re dealing with channel conditions that inherently cannot be controlled. As a result, the only sensible test we could do with iPerf was focus on maximum throughput in the best conditions we could provide. The only thing that this can highlight is the upper bound of efficiency for WiFi due to the carrier sense multiple access scheme in most cases, and rarely detects a whole class of problems that affect user experience on WiFi.

In order to test these things we’ve moved to using a proper testing system that is actually used by at least a few OEMs today, namely Ixia IoT. While we discussed the possibilities for testing, at this time due to the RF isolation chamber used we are limited to AP simulation only, so we can’t properly simulate clients in the channel without restricting ourselves to a single spatial stream for both the AP and client. This wouldn’t be a very useful test if set up in this manner as most devices today that we’re testing have support for two spatial streams, and many routers have three or even four spatial streams at this point.

The first set of results we can talk about that will be of interest is rate vs range. This is a fairly simple test at a conceptual level, as it simply tries to see how well a device can maintain its performance in the face of reducing signal to noise ratio for a given modulation and coding scheme. This is a good high level test of how well a device can maintain a connection as reception degrades. In this test the HTC 10 had an initial RSSI of -28 dBm while the GS7 was at -21 dBm and the iPhone 6s at -22 dBm, which allows us to calculate the path loss and determine the RSSI as a function of the transmit power.

The results of this test are interesting to say the least. Off the bat, every device had different RSSIs measured, so this meant that everything had different levels of path loss. The HTC 10 seemed to have the most path loss, while the Galaxy S7 and iPhone 6s were functionally identical. However it looks like RSSI is really an insufficient metric here because while the iPhone 6s was able to reach maximum throughput using NSS 2 MCS 8, the HTC 10 and Galaxy S7 did its best at NSS 2 MCS 4 or 5. I suspect this may be just due to placement as device positioning strongly affects MIMO as receive-side spatial correlation reduces the gains that MIMO can provide. Regardless, the HTC 10 somehow manages to beat the Galaxy S7 through much of the curve, but for some reason suffers from a reduction in throughput at higher transmit power. It's worth mentioning though that this test doesn't allow for testing of antenna gain or similar tests. Given various levels of futzing about with the device positioning in the test chamber I'm fairly confident that the Galaxy S7 is consistently better with regard to path loss, so even if it doesn't perform as well at a given RSSI it tends to have a higher RSSI than the HTC 10 by about 5 dBm which is fairly significant.

Finally, the other test that we can run at this time is the roaming latency test, which tests how well a device can hop from one access point to another as the received transmit power rises and falls. If you ever rely on WiFi to work as you walk around any building larger than a single apartment unit, you’re going to feel the effects of high roaming latency as VOIP calls or any real-time network application will either experience interruption or drop altogether if roaming is not implemented properly.

WiFi Roam Latency

In the case of the Galaxy S7, roaming latency is honestly rather wanting. In the best case the Galaxy S7's roaming latency appears to be acceptable, but it's still significantly worse than the best we've seen so far. It seems that Samsung's algorithms have issues with edge cases as I've seen multiple instances so far where the device just can't handle roaming consistently. Despite consistent positioning and identical test setup I've seen cases where the Galaxy S7 has problems with consistent roaming. Even with the simple case of 10 dBm to -45 dBm at 3 dBm drop per second, I've encountered weirdness where the device drops from the network altogether claiming that the password given was incorrect (it wasn't) or a few successful handovers followed by getting stuck on a single access point or dropping from the network entirely. Even in the best set of trials performed I still saw 3 of 64 trials fail to roam correctly. The performance is certainly far better than something like the Google Pixel C, but Samsung should really be focusing on improving here.

Video Performance Charge Time and Miscellaneous
Comments Locked

266 Comments

View All Comments

  • LuoSKraD - Tuesday, July 5, 2016 - link

    So you agree because they are bad mouthing a company and a phone you don't like? They didn't cover macro shots or anything at all it was preference. It just seemed like another iPhone biased review, undersaturation poor depth of field, bad low light quality. These are just a few examples where Iphones fail at, this was not a thorough review at all it was just focused on being biased towards making a phone look worse than it actually is. There are much better wrote and explained camera reviews out there for instance.
  • victorson - Tuesday, July 5, 2016 - link

    Don't know about AnandTech, but sure seems like you have a certain bias.
  • retrospooty - Tuesday, July 5, 2016 - link

    I have been coming to Anandtech since 1998, and have always enjoyed it's great reviews... But for the past 3-4 years, they have become very Apple-centric. The comment from LuoSKraD above is pretty correct. It all made sense when Anand himself left the site to go work for Apple, but the clear bias exists, and this review shows it. That is OK, it's not like there aren't other Apple-centric sites out there and it is perfectly withing the sites rights to run thigns the way they want to... But if you are looking for unbiased reviews on any products that compete with Apple, this site is no longer one of them, so browse accordingly.
  • The Garden Variety - Tuesday, July 5, 2016 - link

    Action, man! You're uncovering a great truth here, but we need action! Our opinion matters more than others—how do we make sure our anecdotes and opinions are enforced in all future reviews? We must keep visiting here, so how do we make sure this community only reflects our opinions? I think this is essential and I need your guidance here.
  • retrospooty - Tuesday, July 5, 2016 - link

    I dont know if we can. The site was sold to "Purch" in 2014. The same parten company that owns Toms hardware. LOL. suffice to say they arent interested in the tech, but the clicks.
  • retrospooty - Tuesday, July 5, 2016 - link

    damn typeslexia... Same "parent" company.
  • Meteor2 - Wednesday, July 6, 2016 - link

    You're being trolled :)
  • victorson - Wednesday, July 6, 2016 - link

    AT is one of very few websites that tests phones and bases its opinion on rigorous testing than most reviewers. You are entitled to your opinion about bias (as always, without any factual backing), but it's extremely irritating how all of the popular blogs praised the Galaxy S7's camera without mentioning any of its obvious flaws. That's what I call bias: ignoring the objective reality, so that an opinion of the 'best camera ever' can be manufactured and repeated to infinity. The truth, however, is different: the Galaxy S7 has a very fast camera with amazing auto-focus, but the quality of its images suffers as a result.
  • lilmoe - Wednesday, July 6, 2016 - link

    The dual pixel focusing system alone makes the GS7's camera leaps and bounds better than the competition. If a reviewer ever so slightly undermines this aspect about the GS7's camera (or any camera in that regard), or believes it's just "OK" needs to work more on their photography skills and overall knowledge.

    There are DSLR and mirrorless manufacturers out there that release sequels to their "pro" line cameras with the only significant upgrade being the focusing system, and pro photographers actually upgrade to that new product.

    The focusing system, in short, either makes you take the image, or end up with a blurry shot where no post processing can ever help.

    If you're THAT worried about post processing, then you need to be shooting RAW. I'm perfectly happy with the post processing on my GS7, and when I KNOW I need all the detail I can get (portraits, landscape, sports) then I simply launch pro mode.
  • jospoortvliet - Wednesday, July 6, 2016 - link

    Well, I am fascinated by how different people rate the extreme sharpening Samsung employs. I find it horribly ugly and I keep hoping for either a fix or a setting to disable it on my S7. the HTC one I owned before (m7) made it configurable... please Sammy, copy that.

    I was surprised how undersaturated that night pic was, mine look nice and I agree the fast focus is super important.

    But I do appreciate the test of the radio, that is interesting.

Log in

Don't have an account? Sign up now