Originally Posted by ca4224
Thank you very much - very helpful.
1a) I can suppose that all ports are 60hz but preferably use inputs 1 and 2 then. Great news if they are all the same!
1b) From what I can understand concerning chroma is that 4:4:4 is maximum possible quality being delivered. Now, I've only heard PC gamers concerned around the chroma issues. Is this even an issue when it comes to 4k Blu-Rays, etc. You mentioned Blu-Rays in general.
This is what I read: http://wolfcrow.com/blog/chroma-subs...ers-explained/
So if this TV is 4:2:0 we get 50% of maximum (4:2:2 would be 66%) but this all may be moot because 4k materials will never uncompress to this level.? PC gaming is different then.
2) Good to know and I'm assuming then the difference between a 10 bit and Sony's 8 bit is minimal.
1b) Since all consumer video is encoded 4:2:0 (and I'm talking all DVDs, all BDs, all future 4k BDs, all streaming), that's all that's required for the best quality reproduction. The TV will upsample where necessary, which considering the good processor on the X850B, is probably a better idea than letting the source do it anyways.
However, here's one point that everyone keeps forgetting about
: where are these mysterious 60fps BD videos? Any mainstream movie is going to be released at 23.976fps, and the Sony already supports that at 4k 4:4:4 (not that BDs will *ever* be 4:4:4, as the format specification doesn't even allow it).
But should you find a 4k 60fps video, the Sony will indeed support that if the BD player can output native 4:2:0. I suppose this might be the biggest challenge - finding the proper output support. One can assume that Sony's forthcoming 4k BD players will do this without issue, since they have several years of 4K TVs to support.
So, why might you need 4k 60fps 4:4:4? Primarily, to use the TV as a computer monitor. I'm somewhat doubtful that there are many games with enough fine detail at 4k resolution where you can actually see an obvious difference between 4:4:4 and 4:2:0 encoding, though.
Sure, 4:4:4 support is a nice thing to have, but it's not the be-all that some people make it out to be. Chances are, most folks will never even have a legitimate use for it.
2) Due to how companies market things, quite a few panels are advertised at a certain bit-depth when they're actually one step lower plus dithering. For example, many 6-bit panels were advertised as 8-bit... but with a footnote that said they supported 16.2 million colors - instead of the normal 16.7 million of a true 8-bit panel. The Sony is at least true 8-bit, and I don't doubt that many of the first 10-bit panels marketed will in fact be 8-bit with similar types of dithering. But they'll be "all new" of course! ;-)