I'm using my monitors own colour profile in Windows 10 and wan't to know which AMD driver setting I should use in order to display colors accurately. Within the driver there are two options for color temperature, Automatic and Manual (6500k).
Switching between automatic and 6500k I see a difference in blue and red colours.
My question is which one of those settings is the "correct choice" for accurate color display in my case?
I have set the pixelformat to Full RGB 4:4:4 PC Standard.
Are you sure? Because the default driver setting is 6500k, so when you click on reset it will go to manual.
Can you explain as to why automatic is the right choice? Thx
Thx for the input. Just to get a better understanding, do you by chance know what the manual setting is for then? Why would someone use 6500k over automatic?
If they had no other way to adjust their calibration, such as a laptop, but it's really a holdover from pre Windows 7 days when the display calibration was very limited. Also, if you wanted to temporarily adjust your settings to cheat at games, you could use it.
I recommend 6500K.
HDMI/DP use standard colour spaces (sRGB, BT2020 etc) that typically use 6500K and well-behaved monitors will convert to their own colour temp/gamut. Some do not, they simply reinterpret data in their internal colour space which is wrong.
In those cases Automatic is better - it means the driver will convert to monitor gamut. But if the monitor is well-behaved, then Automatic is counterproductive since both driver and monitor are doing that conversion.
@6500k the colors on my monitor are a bit stronger compared to "auto", but I can't tell if 6500k is oversaturated or natural, because the difference is marginal -but still noticable. On my Samsung 4k TV the setting is called "native" and "auto", and I can tell that "native" is oversaturating the picture there, so I go with auto, which seems more natural. But on my monitor it's hard to tell which setting is the correct one, since both look ok, just that 6500k is slightly more vibrant.
PS: The new AMD driver 19.12.2 changed the menus and now the setting is called "Custom color", and I noticed when the setting is "off", it equals to 6500k.
This must be a bug, because setting "custom color" to "on" and turning off "color temperatur control" will equal to "auto" setting. I think that they made a mistake there.
I have checked it with my colorimeter. Here you have the correct setting:
Custom color disabled= accurate color
Custom color enabled+color temperature control enabled= accurate color
Custom color enabled+color temperature control disabled= innacurate color
It's not a bug, though I admit the naming is confusing.
HDMI and DP have video metadata so GPU can tell the monitor which color space it's using. The options are limited to a few standard color spaces, sRGB, BT601, BT709, BT2020, P3 etc.
Fact 1: It just so happens that all these standard color spaces in PC space use D65 white point (i.e. 6500K color temperature).
In 19.12.2, "Custom Color" disabled means that color temperature is matching the color temperature of the color space indicated in the video metadata. This is really how it's supposed to work - HDMI/DP use some standard color spaces, GPU tells monitors which one it's using, and monitors convert to their native color spaces.
When you enable "Custom Color", default setting is Color Temperature Control enabled with 6500K. Due to Fact 1, this is the same as if you disabled "Custom Color" altogether. This is why it's not a bug, using 6500K is the right default.
With Custom Color enabled, and Color Temperature Control disabled, GPU is doing the conversion to panel's native color space. It's really not obvious from the name, it used to be called differently.
The problem is that monitor doesn't know this, and if it follows HDMI/DP spec and honors video metadata (this is what I meant by well-behaved in my earlier reply), it will convert from the color space in the video metadata to its native color space, without being aware that GPU already did that. Basically, this setting will only work if the monitor is ignoring HDMI/DP spec regarding color space info. Or maybe with DVI where there's no metadata.
This is why in most cases "Color Temperature Control disabled" isn't right.