Color Space

A color space is a subset of entire visible light spectrum, and the reason that it’s a subset is that being able to reproduce the entire visible spectrum is technologically impossible at this time. Various color spaces will all target a different subset of colors, although they may overlap with each other. A color space is also known as a color gamut.

The most common color space on a PC today is sRGB, and almost all laptops and monitors target this color range. But in the past several years, we’ve seen more and more devices able to target a gamut with a wider range of colors, such as Adobe RGB, and there are advantages and disadvantages to using these wider gamuts on a PC.


The sRGB Color Space - Source: SpectraCal

The sRGB, or standard Red Green Blue gamut is an RGB color space which is the standard used on all PCs and the Internet. If you look at an image on the internet, it’s more than likely an image within the sRGB color space. Windows defaults to using the sRGB color space for everything.

Other common gamuts used in the PC industry are the Adobe RGB gamut, developed by Adobe in 1998, and the P3 color space, both of which are wider than sRGB. Wider means that they cover a range of colors larger than the sRGB gamut.


The Adobe RGB Color Space - Source: SpectraCal


The P3 D65 Color Space - Source: SpectraCal


The BT.2020 HDR Color Space - Source: SpectraCal

But being wider isn’t always better, especially when using Windows as your operating system. Windows defaults to sRGB for pretty much everything, and doesn’t really have any sort of robust color management system (CMS) that can transform colors targeting one space into another. Applications can build and use their own CMS, but that’s at the application level only, and requires the developer to do much of the legwork. If you think of an sRGB color represented to be displayed on an 8-bit display, if you want to display 100% red you’d use 8,0,0 on that pixel so that red is on the maximum brightness (typically 255 or 1.0f). But if Windows opens a photo that was created as sRGB on a display that covers the P3 gamut, it will still tell the display to use 100% red, zero green, zero blue. But 100% red in sRGB is closer to 80% red in a P3 display, so the colors will be overblown, and the image won’t look correct.

Incorrect color on the Left (Photos) vs correct color on the Right (Adobe Photoshop Elements)

To Microsoft's credit, Windows is getting better. But progress here is slow due to legacy matters and the need to avoid breaking existing software, so using wider color spaces on the desktop is still a dicey proposition.

The end result is that on a laptop, offering a wider gamut display can cause problems for most workloads. And for that reason, unless you do a lot of work in a color managed application like something from Adobe, it’s generally not ideal, unless your laptop has the ability to switch the color gamut the display uses through software.

Finally, there are even wider gamuts available, like the Rec. 2020 color space. But at the moment there’s no technology available in a laptop-sized implementation that can create all of the colors in that space.

White Point

The white point on a PC display is generally going to be D65, which corresponds roughly to sunlight at midday in western or northern Europe. When discussing the color temperature in Kelvin, D65 is close to 6504 K. When discussing a DCI-P3 display, the target white point is D63, which is more green than D65. In the PC space, you’ll generally be dealing with a P3 D65 gamut, rather than DCI-P3 which is used in the cinema. Though there are a few PCs, such as the Surface Studio, that can target either DCI-P3 or P3 D65.

How many bits?

Most displays are 8 bits-per-channel, meaning each red, green, and blue channel can have 256 (2^8) steps, which provides 16,777,216 (256 * 256 * 256) different color combinations. That sounds like a lot of colors, and it is, but on each channel there’s still a fairly large jump between steps, since red can only be 0-255, for example.

Less expensive displays may even reduce this more to 6-bit with Frame Rate Control (FRC) which uses temporal dithering which quickly switches between two colors to simulate the color in between to simulate the full 8-bit levels. Often TN displays are only 6-bit with FRC, and lower priced IPS such as e-IPS can also be 6-bit.

Meanwhile with the wider color gamuts available now, the jump in a color step for each increase from 0-255 can be even wider. So lower-bit panels are more susceptible to color banding on the display, which as the name implies, is a situation where distinct bands of colors appear when the intention is for a smooth gradient. This occurs when two colors next to each other are supposed to be slightly different, but end up looking exactly the same or more different than intended, since the display can’t create the color in-between. To counteract this, more bits are needed to represent smaller graduations from one step to the next, so often a wider gamut display will be 10-bit, or 8-bit with FRC. A true 10-bit display offers 1024 levels for each channel, which creates 1,073,741,824 different color options.

Although all modern GPUs support 10-bit, NVIDIA currently restricts 10-bit on OpenGL to their Quadro lineup, so applications using OpenGL like Adobe’s suite require a professional GPU.

High Dynamic Range

High dynamic range, or HDR, is one of the best new features to come to displays in quite some time. HDR brings a wider range of brightness levels, and requires a higher bit-depth as well, to keep the detail available in darker scenes. To provide that, displays certified for HDR have to achieve higher brightness levels than traditional displays as well, so it can be a large benefit.

Source: Samsung

VESA has announced the DisplayHDR specification, which has three levels: DisplayHDR 400, 600, and 1000, which each level corresponding to the peak brightness of the display in nits.

There are two main competing HDR transport standards right now, although only one is available on the PC. HDR10 is a 10-bit format with static metadata for the HDR info, and Dolby Laboratories Dolby Vision is a 12-bit color depth with dynamic metadata which allows the mapping of the pixel data to the luminance to be adjusted frame by frame. PCs only support HDR10, and not only do you need an HDR monitor, you also need at least a Kaby Lake iGPU or newer, an NVIDIA 900 or newer series GPU, or AMD Radeon RX 400 series or higher, as well as DisplayPort 1.4 or HDMI 2.0.

There are only a couple of laptops on the market which offer an HDR display at the moment, but with the eDP connector and complete control over the product stack, HDR in a laptop is likely to be easier for the consumer to use. As of WIndows 10 version 1803, the built-in display needs to have a resolution of 1920x1080 or higher, with a recommended brightness of 300 nits or more. The device has to have integrated graphics with PlayReady hardware DRM for protected HDR, and of course the correct codecs for video which is generally going to be HEVC. Kaby Lake or newer support this. The nice thing will be consumers won't have to worry about any of this, since it will be up to the OEM to ensure everything is in order, rather than with a desktop system where the user has to make sure all of the i's are dotted and the t's are crossed to enable HDR.

There’s a lot of new HDR products coming to market soon, and when it becomes mainstream it’ll be a welcome benefit.

Building the Transistors and Lighting the Display How We Test Displays
Comments Locked

49 Comments

View All Comments

  • heffeque - Tuesday, July 10, 2018 - link

    Since recently, for me an important part of the decision regarding buying a laptop is PWM.

    When I start getting a headache I tend to lower the brightness to make it easier for my eyes, and I recently noticed that lowering the brightness was actually making it worse. I read a few articles and saw that it was due to PWM. Since then, if I start getting a headache, I put the brightness at 100% and it goes away. So I'm definitely going to be more careful choosing my next laptop and/or screen regarding the PWM.
  • nfriedly - Tuesday, July 10, 2018 - link

    Why don't they make RGB LED displays? That seems like it would have a lot of the benefits of OLED, but without the risk of burn-in.

    Also, you mentioned RGB LED backlighting - how does that work? What's the benefit?
  • DanNeely - Tuesday, July 10, 2018 - link

    We can't make normal LEDs small enough for individual pixels unless you're looking at something like a Jumbotron sized display.

    LED's small enough to serve as pixels for a monitor are aspirationally branded as micro-LED, and are a subject of active research. To get an idea for where we're at with something near production ready, there's talk that the 27/32" gen 2 FALD LCDs will have an ~1000 zone microLED backlight - up from the current 384 zone models.

    You can get a wider range of colors out of separate red, blue, and green LED backlights than from a single white LED. The tradeoff is that they're more expensive and draw more power.
  • Brett Howse - Tuesday, July 10, 2018 - link

    What Dan said :)
  • boeush - Wednesday, July 11, 2018 - link

    I'm confused about the power math. I'd think one big LED (let's say 1 cm squared) would consume the same amount of power outputting the same amount of light, as the same LED broken into smaller pieces (let's say 100 mm-squared chunks) - or am I wrong? I thought for a given LED technology, you get X units of light per Y units of energy per unit of area - i.e. I thought efficiency of an LED does not depend on its dimensions but only on its materials, down to some quantum-level limit - am I wrong?
  • DanNeely - Wednesday, July 11, 2018 - link

    The wider gamut means you're putting light into a larger section of the visible spectrum. You've also got 3x the control circuitry/etc in the loop drawing power.
  • MrSpadge - Wednesday, July 11, 2018 - link

    What Dan said. Plus: if you distribute the LED area, it will run cooler and gain some efficiency. But the comment about increased power draw was referring to the RGB backlight, not to distributing micro LEDs.
  • MrSpadge - Wednesday, July 11, 2018 - link

    "We can't make normal LEDs small enough for individual pixels"

    It's actually relatively easy, at first thought: use the substrate as common back contact, epitaxially grow the emitting layer and use a decently transparent conductive oxide as top electrode, just like in OLEDs. Structure the top contact into pixels.

    However, doing that is simply prohibitively expensive as the entire display area would have to receive the super expensive expitaxial growth process, which generally only works on small samples (2 - 4") and requires expensive substrates. Regular LEDs are only affordable because the amount of actual chip area per device is tiny.

    And with a common back gate there's be a major problem of adressing many pixels individually, so one would need some trick to structure the back contact as well.
  • mr_tawan - Friday, July 13, 2018 - link

    I think Samsung currently has a 146" micro-led display demoed already. That screen is 4K resolution, so I guess they could do 75" FULL HD with the same technology.

    However even at 75", it's still too large to my liking :)
  • Martin654 - Wednesday, July 11, 2018 - link

    When looking at the video for a laptop there are four items to look over: screen size, resolution, screen type and graphics processor. For most people, only the screen size and resolution are all that will really matter. The graphics processor really only tends to make a difference for those looking to possibly do some mobile gaming or high-definition video but they can be used for more than that. Pretty much all laptops use some form of backlit active matrix display to allow for bright fast displays capable of video playback.

Log in

Don't have an account? Sign up now