Jump to content

cpc

Members
  • Posts

    210
  • Joined

  • Last visited

Posts posted by cpc

  1. ​Thanks for the great feedback! I actually did an Autumn Leaves test a few weeks back and found the colours amazing but it was very, very contrasty so I didn't do anything with it. I got the impression that dropping contrast to -3 wouldn't really help but might give it another look. I do love the flatter look you get with SLOG2.
    Will also take your tip with the skies and make them a little punchier, it's also very teal-y at points but that might be the LUT at work.

    ​You can increase the DRO setting if you are going to use Autumn Leaves for grading. DRO +3 flattens the low-mid range of all the styles.

  2. It doesn't have more highlights, whait it does is pull mid grey lower and clip signal at IRE 100. The lower midgrey point may create the illusion of more highlights latitude, but it is the actually opposite. The other three Cine curves go to IRE 109. Cine2 is supposed to be used "as is" and be broadcast safe (hence clipped at IRE 100).

    I don't think any peaking setting will work well with slog. Peaking relies on contrast and contrast is very low with s-log. One way to sort of work around this is to pre-focus with a contrasty PP and switch to s-log for recording.

  3. Log does work very very well with 8bit. 8bit is not the issue, weak codecs are. 

    Example, Canon C300, 8bit 50mpbs, Canon Log works a treat, even with the 8bit 24mbps C100, works beautifully. Sony A7s, works great (no banding or codec break up when while de-logging - some poeple don't like the a7s colour science but that's irrelevant).

    Log is simply a logarithmic gamma curve to get a flat tonal range easier to grade and protect both ends of the spectrum it will work with any codec that has the ability to be strongly contrast-adjusted without introducing artefacts. 

    GH4 4K 100mbps codec is one of the strongest 8bit codecs, as is the 200mbps 1080p, not to mention the ability to record 10bit 600mbps 1080p ProRes HQ through HDMI (You can get that for 300$ Ninja Star or 500$ Ninja2 that gives you an extra 5" field monitor)

     A strong Log curve coupled with tweaked colour science (which is what V-log is, it's also a little colour magic) is a perfect companion for the GH4. 

    ​Well, I have the A7s and s-log2 is unusable in 8-bit if you care one bit about image quality. Plastic skin everywhere once you restore contrast; simply not enough tonal precision in 8 bits for the flatness of this curve. Besides, the s-log2 setting doesn't use a huge chunk of the coding space, this way further screwing the tonal precision.

    Canon c-log is a different story. Its curve is not log really, just log-like in the upper end. And it has less DR than s-log. Much more 8-bit friendly.

     

    Codec is important but not as important as the bitdepth. Try working with 16 stops log in uncompressed 8-bit and see what happens even if there is no codec at play.

     

     

  4. This is probably for historical reasons and is related to interlaced video.

    With interlaced, each of the two fields is subsampled separately (because the fields represent different time moments and subsampling them the same way as with progressive images would introduce chroma artifacts related to motion). Now, because each field is subsampled separately, if you use one of the subsampling methods with only half the vertical resolution, for example 4:2:0 or 4:4:0, this would result in gaps of two lines with no chroma samples.

    Here is how a column of 4 neighbor pixels looks like in this case:

    Field 1 Top row (chroma sample)

    Field 2 Top row (chroma sample)

    Field 1 Bottom row (no chroma sample)

    Field 2 Bottom row (no chroma sample)

     

    But if you use full vertical sampling, as in 4:2:2, there is no such issue. You only get 1 sample gaps horizontally, and no gaps vertically when applying 4:2:2 to interlaced video.

  5. Well, since you've been renting lights you should know best what works for you?

    I'm not sure how the 1x1 lights can be enough for lighting a normal scale set/location from scratch. Are you only shooting closeups or using the LEDs to augment daylight?

  6. 1) Convert your ML raw/mlv files to dng sequences using one of the many available GUI batch tools. This is essentially a simple and fast rewrapping job. 

    2) Delete the raw/mlv files. Raw process, edit, color and finish in Resolve Lite, non-destructively, using the dng files. 

    No generations, proxies or whatever. Full original quality at each step. That's what I've used. Now, if you need compositing you'll have to export to a compositing app using whatever format suits you. I've used tiff sequences for that.

    (I have an extra step, using my compressor http://www.slimraw.com/ to losslessly compress the DNGs (usually 50+% size reduction while retaining full original quality), but this is optional. /end of plug)

     

    And apparently there are now ways to mount the ML raw files directly so that they are visible as dng sequences in video apps. You should consult the magic lantern forum. Tons of workflows there.

    I haven't tried CC 2015 yet, but it looks promising. CC 2014 was way too slow for my liking when working with raw.

  7. ​What I would like to know is whether there is agreement on how much colour depth is gained when downsampling 8-bit 4K to 1080p. Is it comparable to, say, the 10-bit 422 that the BMPCC outputs?

    Well, I wrote this last year: ​http://www.shutterangle.com/2014/shooting-4k-video-for-2k-delivery-bitdepth-advantage/

    Downscaling 4:2:0 8-bit video from 4K to 2K will give you 4:4:4 video with 10-bit luma and 8-bit chroma. But have in mind that these are the theoretical limits. In practice, what you gain depends on pixel variation and compression used on the source image. The more detailed the 4K image, the more true color precision you gain in the downscaled image.

    In any case, downscaling 4K in post delivers the best looking 2K/1080p from the current gen 4K cameras.

  8. Looks good.

    I would probably have him turned a little more to the left and also move the camera a bit left to have the window offset, and his head against the blue columns. Also, the light would probably be slightly better on him. The changing sun is messing a bit with the brightness of these interview shots, they go continuously darker. Not a big deal, but you may want to match them.

  9. You are downscaling 4k to 1080p, Then upscaling 1080p up to compare to 4k. Downscaling is not a reversible operation. Why? Because different groupings of pixels can result in the same output pixel once downscaled. When upscaling back the software can't create the original pixels out of thin air, because there are many possibilities for these pixels. In the end, you are comparing a 1080p image to a 4K image. And of course, the 4K image will have more resolution.

    So a proper downscaling is losing you resolution and gaining you bitdepth/color precision at the lower resolution, in the sense that each new pixel is essentially "quantized" at a higher bitdepth.

  10. cpc- I've sharpened the 4K GH4 footage in post and the noise grain is very fine. There are smeared areas and macroblock artifacts in some places, though overall when there's not a lot of motion the noise grain is pretty impressively small, especially compared to my 5D3 (RAW) or FS700 (AVCHD).

     

    Premiere uses a form of Lanczos and Bicubic for scaling ( http://blogs.adobe.com/premierepro/2010/10/scaling-in-premiere-pro-cs5.html ). Whatever they are doing appears to work reasonably well. In terms of the 10-bit luma debate, if the 8-bit 4K footage was effectively dithered, either via error diffusion or simply noise, then resampling a 4x4 grid  (Bicubic) could kind of reverse the error diffusion and put some bits back into luma. Intuitively it doesn't seem like it would buy a lot of latitude vs. native 10-bit sampling, however a little bit can be helpful. Adding error diffusion / noise certainly helps reduce the appearance of banding/blocking. Ideally the dither/noise would only be added where it's needed. Without significant dithering of some form or another, I don't see how 4K to 2K could do much for the '10-bit luma' argument as we need variance for the 4 source pixels to spread around the values of the summed/averaged final pixels.

     

    But there is usually some variance, moreso with multiple samples per output pixel. Yes, noise provides this, but so does detail and even gradients can have it (the steeper, the better), especially without compression on top.

    The thing is, in blacks and dark grays, where precision is most lacking, noise is at its strongest. So in practice you gain the most exactly where you need the gain.

    (On a remotely related note: Been playing with some Kinefinity footage lately; noise is nice and the image scales down to 2K beautifully.)

     

    @Axel: Not necessarily the case with HDMI out being subsampled the same as the in-camera recorded video. Lots of cameras record 4:2:0, but output 4:2:2 on HDMI. And some go 4:4:4 (over SDI), F3 comes to mind.

  11. Not sure about that, John. The native fine noise in the 4K image is probably mostly killed by compression as much as the downsampled grain in the original 2K image is killed by compression.

    I don't think NLEs dither, Resolve doesn't seem to do it. But the scaling algorithms are surely involved, and this affects output because quite a lot source pixels are sampled per output pixel (a generic cubic spline filter would sample 16 source pixels).

     

    @sunyata: it is not true 10-bit, and in your example chroma is still 8-bit (with no subsampling though), but it surely is better than a 2K 8-bit subsampled source from camera, if not as good as a true 10-bit source.

  12. Thanks to John for pointing me here, it is an interesting discussion. :)

     

    As one of the people who think that there is free tonal precision lunch to be had in 4K to 2K downscale and the one that wrote the ShutterAngle article linked on the previous page, I think the 8-bit display argument is a bit beside the point. The whole idea of shooting 4K for 2K (for me, at least) is in using a flat profile as s-log2 on the A7S. Then working it in post to the appropriate contrast. 

     

    As my idea of a good looking image is generally inspired by film and includes strong fat mids and nice contrast, this means the source image is gonna take quite the beating before getting in the place I want it. And here is where the increased precision is going to help. To simplify it a bit: when you stretch an 8-bit image on an 8-bit display, you are effectively looking at a, say, 6-7-bit image on an 8-bit display, depending on how flat the source image is. That's why starting with more precision is helpful. Starting with 20-25 values in the mids (which is the case with 8-bit s-log2) is just not gonna handle it, when you are aiming at, say, 60-65 values there in delivery.

     

    Compression and dirty quantization to begin with surely affect the result and limit precison gains. But they don't entirely cancel them, and the better codec you use on the hdmi feed, the cleaner the downscale.

  13. My guess is the FC process goes like this:

    1) Linearize source data. Source can be any transfer curve, and with varying colorimetry. Tonal curve is linearized, and color is transformed (corrected) to some reference color space. This step equalizes the input.

    2) An idealized "printed" film transfer function (with the film negative gamma/contrast index expanded to 1, hence "printed") is applied, which pushes colors around, possibly also tweaking contrast (based on the film negative contrast index values).

    3) Result is gamma encoded for display.

     

    In theory, 1) and 2)  (well, and 3, for that matter) can be done in a single composite step (a composite LUT for each possible source type and each possible target film, for example).

  14. Yes, but that's not what I am asking. Any negative film is extremely low contrast. We are talking film gammas like 0.5-0.6. The image is never meant to be used at this contrast level. The audience never sees the negative. And printing on paper or on release film restores contrast.

    Even higher contrast release stocks help battle theater projection issues (stray light, low theater screen luminance, lateral eye adaptation) which tend to make darks appear brighter. The gamma of the whole system, from scene to projection, is higher than 1, often higher than 1.5, depending on the release stock.

     

    Hence, the question about FC. There is an abundance of FC footage floating the internets which looks unnaturally flat (apparently, for many people lifted blacks = filmlike), and definitely not in the way a specific stock really looks like when printed. I doubt that FC defaults to digital cinema contrast - digital cinema gamma curves actually have higher contrast than computer displays (sRGB, 2.2), exatly due to the projection issues mentioned above.

×
×
  • Create New...