Blog Show [SOLVED] TN 4k monitor = Display Colors : 1073.7M (10bit) IPS 2k monitor = Display Colors : 16.7M (real 8 bit)
Status Not open for further replies.
I was using TN panel 1ms responsetime 4k (3840x2160) 60Hz monitor for 2 years. The most diffrence for me is the colors: The 4k monitor = Display Colors : 1073.7M (10bit) For me thats a huge diffrence in color space,
or maybe its just the IPS and TN look complelty diffrent? 144Hz feels smooth and good to play but i miss the colors from TN panel. How do i make the IPS looking good? its all on default settings now. TN panel 4k monitor = Display Colors : 1073.7M (10bit) 8bit vs 10bit = 1073 milion vs 16 milion If there is a way to make IPS colors look good id be happy if anyone can explain in comments how to do it. i got no clue^^ Last edited: Nov 25, 2019 Apr 15, 2014 11,053 1,216 54,340 1,374
FYI, just because the monitor is capable of displaying 10 bit color doesn't mean any of the content you're looking at is in 10 bit color. Try changing between 10 bit and 8 bit on the 4K monitor, I doubt you'll be able to see a difference. Which exact monitors? Edit: You can actually have 10 bit monitors with poorer color gamuts than an 8 bit monitor. IPS panels usually have
better colors than TN. Apr 15, 2014 11,053
1,216 54,340 1,374
FYI, just because the monitor is capable of displaying 10 bit color doesn't mean any of the content you're looking at is in 10 bit color. Try changing between 10 bit and 8 bit on the 4K monitor, I doubt you'll be able to see a difference. Which exact monitors? Edit: You can actually have 10 bit monitors with poorer color gamuts than an 8 bit monitor. IPS panels usually have
better colors than TN.
Aug 8, 2015 2,960 271 25,990 824
Higher color depth (bits) doesn't improve color space coverage, it just allows for finer gradients. 8 bpc color depth is already sufficient, 10 bpc color depth provides almost no visible improvement. Status Not open for further replies.
2021-07-23 at 17:12 #31052 After getting a new monitor few days ago – ASUS PG329Q (10 bit + gaming) – I started my journey of calibrating a wide gamut monitor (previously only did it for 8 bit) and trying to understand what owning a 10 bit monitor really means. I have run numerous calibrations, using DisplayCAL including 8 bpc and 10 bpc settings in nvidia control panel. I am a bit surprised that there is a quite negligeble difference in results – at least to my understanding – the percentage of coverage of sRGB, Adobe RGB and DCI-P3 are nearly identical for 8 and 10 bpc. Is this expected or am I doing something wrong? My results after calibration are at best like this for gamut coverage:
Gamut volume is at 180%, 124% and 128% respectively. ASUS advertises 160% of sRGB colors and 96% of DCI-P3. Does having a 10 bit monitor make any difference for calibration result numbers? Should I be using separate ICM profiles for calibration at 8 and 10 bpc, depending in which mode I am running (this changes based on refresh rate of the monitor, only 60 Hz so far works with 10 bpc, while I run games with 165 Hz).
2021-07-23 at 18:12 #31059
Expected. Coverage is given by LED backlight spectral power distribution, not by panel.
With dithered ouput at app or at GPU HW output no real difference. Without them, if monitor accepts 10bit input, even if panel is 8bit, you can use some app lack of features to let monitor handle the rounding error instead of the application, because if aplication goes down from 10 to 8 without dither before sending to GPU likely to cause some truncation errors. People and vendors too usually mix “accepts 10bit input” with “10bit input at panel” with “true 10bit panel”. They are different depending on who has the responsibility to truncate : app, monitor HW, monitor panel… although if properly done results are interchangeable on SDR contrast windows (256 step can cover that kind of window with dithering).
2021-07-23 at 18:28 #31061 Simply draw grey (black to white) gradient in Photoshop, you’ll see it. 2021-07-23 at 18:36 #31063 From your comment I understand that there can be 3 cases of different monitor hardware: “accepts 10bit input” with “10bit input at panel” with “true 10bit panel”. For my model, the only things I can tell is that:
The specs can be found here: https://rog.asus.com/monitors/32-to-34-inches/rog-swift-pg329q-model/spec However, my question was more general about any 10 bit monitor vs 8 bit. My limited understand of the topic and your comment is that there is no difference in color qualities (?). Please correct me if I am wrong here. 2021-07-23 at 22:43 #31068
Because wrong truncation in PS, not because it was needed.
It is not a partition. A device can have 1 or 3 features. Asus claims are at least “accepts 10bit” which is the requirement of poor implementation in Photoshop to avoid truncation on 16bit gradienst since PS is not capable of dithering. LR or C1 do and do not need nor use 10bit.. The funny part is that photographers do not need it (10bit output) and designers and illustrators who are likely to work with synthetic gradients cannot use it because Adobe has not (AFAIK) 10bit output or dithered output for them.
For SDR contrast window no, there is not, unless poor output to screen like Gimp, PS, Ai, In…. 2021-07-24 at 9:54 #31071
Switching to Microsoft ICM you get some cleaner grey, but tints are still visible. You also see them in browsers and viewers, this is general problem in Windows. But, being plugged in Macbook and calibrated, 8-bit display shows clean grey. Here is the same thing, Vincent: we may talk on theory and tech aspects, but 10-bit gives practical advantage to Windows users by now. Opinion: I show this effect to photographers and describe how to check gradients purity, – totally switch off ICC usage in two steps: flash vcgt in Profile Loader and use “Monitor RGB” proof. It’s important for B&W and mixed studio shots, commercial design and design over photo (popular in product photography) as well. This take no sense in real world colour photos. 2021-07-24 at 13:01 #31072
No, that is false, it is not monitor, it is color management what causes banding. Do it with dither and there is no banding (ACR,LR,C1, DMWLUT, madVR..)…. unless GPU calibration causes it. This last one is NOT windows related, it is related to HW in GPU. AMD can dither on 1D LUT, even on DVI connections, other vendro may fail (intel) or hit & miss (nvidia registry hack, here in this forum there was a thread). Also
your own post is a proof that you are wrong. 8bit macbook can render smooth gradients in PS because Apple provided an OpenGL driver that have a “server hook” at 10bit to client app (PS), then driver do whatever it wants, dither to 8 or send 10bpc if chain allows it: the kay is that poor PS implementation regarding truncation was avoided. Actually if every GPU vendor provide that hook avery display even on 8bit DVI link can show bandless color managed gradients. Nvidia does for gamer GPU (studio driver) although 1DLUT can be problematic, newer AMDs can enable it and also do 1D LUT dither since 10 yers or more. For intel unless Apple custom drivers there is no 10bit hook to app and ther is no 1D LUT dither (but you can use DWM LUT on win). Also color management with 3xTRC and app using 8bit rounding like Firefox is prone to that kind of “color” banding instead of typical grey step banding with 1xTRC. All of this is NOT monitor related at all, just software and GPU HW limitations. 2021-07-24 at 13:46 #31075 I appreciate your deep input on the topic. Could you (both) suggest your recommended monitor specs for photo editing and viewing, primarily on Windows? Apps: Capture One, DxO PhotoLab, Affinity Photo. GPU: Nvidia RTX 3080. Size: 32 inch (my current target) Panel technology: Panel color depth: Backlight technology: Color Space: etc… Knowing these paramters from the experts, I will try to combine them with gaming specs. I know this will most certainly result in some compromises, but I would like to get at least 80% of the way in both aspects. 2021-07-25 at 19:18 #31077 OK, Vincent, but what do you think on signal type syncronization? I don’t mean composite, YCbCr etc., but the effect of data type and derivative adaption of hardware workflow. By example, I have already described the flaw with weak black at some horizontal frequencies. Could similar thing happen with vcgt (2provanguard: videocard gamma table) dithering on/off/level? My experience tells me that 10bit displays realy draw better grey in Photoshop and this happens even with nVidia cards, though 10bit displays are seldom items here. Hm, why do you call it dithering? I know the thread with nVidia hack, but is the effect described anywhere for programmers? 2provanguard: 32″ displays are rare birds in my practice. Mostly they’re of MVA type or consumer-graded IPS. I may only recommend you to find at least two good tests of some model, the most clear testing bench for graphics is prad.de. Asus gamers displays I’ve met are totally ugly toys. They have graphics displays, but these are also strange, don’t trust in their calibration and quality. MSI makes some better monitors, but one of MSI notebooks had terrible color flaw in “pro” software (RGB pallete drop out). You should simply understand that gaming is agressive show biz, so gaming hardware manufacturers won’t care of natural vision. I’ve even met top nVidia gaming card without vcgt at one of outputs. Note that games don’t use ICC profiles as they slowing computations, video editors usually work with LUTs instead of ICC profiles. nVidia bad vcgt may also be a back side of high velocity. Probably some expensive displays combine pretty fast panel (you knows it better), wide gamut (full coverage of some stadard profiles ), correct RGB primaries with separate RGB spectra – better color stablility under different light (too difficult to novices), good uniformity (delta C < 1,5 at square part of the whole screen), high enough contrast (>1200:1 for IPS, but video editing needs more, MVA hass up to 5500:1) and smooth gradiends (check by eye), check also for color-to-brightness stability (avoid jumping colour change).
2021-07-25 at 22:28 #31079
It is no display related as I said. It’s because the whole chain: Photoshop: processing (GPU basic vs accel) -> truncation to interface driver -> openGL vendor driver -> (1) LUT -> output (dither/no dither) -> phys connection -> display input (8/10)-> monitor HW calibration/factory calibration/calibration with OSD) -> dithering to panel input -> panel input (8/10) -> (optional dither) -> actual panel bits. LR/ACR on Photoshop/C1: Processing -> truncation with temp dithering -> Windows composition at 8bit -> (1) GIMP / In design / Ai / firefox /others: Procesing -> truncation to win composition iface -> (1) If display has no banding non color managed, color managed banding is ONLY caused by steps before (1). Photshop chose to do it in the “expensive way” (before gamer Geforces and Radeon), requiring 10bit hook opn OpenGL and 10bit end to end pipeline because GPU vendor needs, before taht “casual 10bit driver” people had to pay for Quadros and Firepros for task that do not require such high bitdepth end to end (others do, but no photo SDR work). Apple know the trick and with RGB8888 (I do not remember name) pixel format they provide hook for 10bit input, although they will truncate with temp dithering on GPU, out of PS scope Also Adobe for other tools chose to do it the RIGHT WAY: processing output dithering to whatever windows composition it has. 0 banding if non color managed the monitor have no banding. Also Adobe for other tools chose to do nothing: truncate to win composition to 8bit: Illustrator/Indesign, which is a shame because syntehics gradienta are common tools there. If you have issues with PS is and none if this is related to 10bit advantage end to end on SDR contrast windows.
Temporal dithering. It works auto in AMD cards (related to 1DLUT output) and in ACR/LR/C1. It is done by default, user need to do nothing.
As said by Alexei IPS/VA & good uniformity. Since you have a newer GPU model you can get one with 10bit input (whatever panel it has behind) so only for Photoshop you can get rid of colormanagement simplifications done in that app (truncation of processing to driver interface with no temporal dithering). Also
since you want a gamer display those new 165Hz 27″ QHD or UHD are usually P3 displays, some of them do not have gamut emulation capabilities so for gamer all will look “wrong” oversaturated… but you can look at @LeDoge DWM LUT app , works like a charm.
2021-07-25 at 22:43 #31081 Hi @Vincent I downloaded the app. Looks very promising! How do I do the following in DisplayCAL: “Use DisplayCAL or similar to generate the 65x65x65 .cube LUT files you want to apply” ? Is using this app effectively replacing usage of ICC profiles (in some situations) ? 2021-07-26 at 0:09 #31082 ICC with GPU calibration and DMW LUT “can be” mutually exclusive, depending on VCGT. Use DisplayCAL and calibrate display at native gamut to your desired white. Install it & etc. Then if you wish a LUT3D for DMW LUT: If you wish to calibrate grey using DWM LUT because your card don’t dither or do not make it properly or because you want to, “apply VCGT” to LUT3D when you create it. BUT if you do this you cant have displaycal profile as display profile in OS. Assign as default profile in PS the ICC of colospace to be simulated. It’s like Resolve LUT3D for GUI monitors, you have to choose who is going to calibrate grey, 1DLUT in GPU HW or LUT3D in software. For non color managed apps if you rely on ICC to gray calibration, no need to change it on OS, LUT3D won’t have VCGT applied.
2021-07-26 at 20:57 #31091
Check!
Source Profile: sRGB IEC1966-2.1 (Equivalent….) Tone curve: Gamma 2.2, Relative, black output offset: 100% Destination Profile: the profile I created using DisplayCAL for my monitor D65, gamma 2.2 Apply Calibration (vcgt): UNCHECKED (Is this correct?) Rest are default values.
Well, I do have the profile created using DisplayCAL loaded and active, so I did not use “apply VCGT”.
Is this only specific to PhotoShop?
From here forward I am a bit lost. How can I rely on ICC if an app is not color managed?
I have noticed that my settings (DisplayCAL produced ICC is loaded) Capture One and DxO Photo lab receive desaturation when I activate the 3D LUT. Burning reds go to reds, and then go to brown with 3D LUT enabled.
I would like to try that. I would require a noob level instruction, please 🙂 2021-07-26 at 21:58 #31092
If your GPU causes banding you can check use VCGT, and assign as default display profile a synth version without VCGT. Explained bellow
I meant “OS” not “PS”, control panel\ color management, PS is a typo :D. To not mess with Photoshop color options if you do not know what you are doing.
VCGT is grey calibration, embebed into disoplaycal ICCand loaded into GPU.
Go to displaycal folder, open synth profile editor. Make a synth profile with the same white, and red, green and blue primaries coordinates (illuminnst relative xyY data on profile info in displaycal) and same nominal gamma. Usually you want to play teh “infinite contrast tick” (black point compensation) on both profiles. Then make a LUT3D with that sytn profile as source colorspace, target your displaycal profile with VCGT caibration. Assign synth profile as default display profile in OS (control panel, color managemen , device tab). Open DWMLUT and load LUT3D. This way you can get no banding even with intel iGPUs, unless VCGT to be applied is to extreme to be simulated with 65 node per color ramp. Games will look oversaturaed (native gamut) but for PS or LR os like yu had an Eizo CS with HW calibrayion and idealized ICC profile (matrix 1xTRC) than minimized banding caused BY color management app.
2021-07-26 at 22:44 #31094
OK. I think I did it. But cannot be sure if I understood all the steps required. Man, if this combination of DWMLUT and DisplayCAL can make my wide color gamut monitor show proper colors on Windows in different apps and games, then this is GOLD! I could easily pay some money for a comprehensive guide on what to do and why, or for further development of DisplayCAL to do the proper things automatically for me. From what I understand I will need to switch different profiles (OS + DWM LUT) for when I use Photo apps and for when I run games or browsers. This could be further automated. Same with generation of the synthetic profile from the ICM profile. What is true 10True 10-bit displays have 10 bits per pixel, for 1024 shades of RGB color (Red, Green, Blue). Cube that and you get to those 1.07 billion shades. Expect smooth gradations and improved realism in photography and video content.
Is 10If you're a Filmmaker or Photographer, then you should 100% get a 10-Bit monitor. Most cameras shoot 10-Bit or higher now, especially Photography DSLR cameras averaging around 12-Bit.
Are there any 10BenQ MOBIUZ EX3210U 32” 4K IPS Gaming Monitor, 144Hz 1ms, HDR600, True 10-bit, HDMI 2.1 (48Gbps), 98% P3, 99% Adobe RGB, FreeSync Premium Pro, Eye-Care, Microphone, Built-in Speaker, Remote Control.
Is 8bit vs 10bit noticeable?The main difference between an 8-bit and 10-bit panel is the number of colors they can produce; an 8-bit panel can display 16.7 million colors, while a 10-bit panel can display 1.07 billion.
|