Most old workflows in the times of analog film, Telecine and analog/digital tape formats were “Display Referred“.

This means the material was seen and graded for the display that you use.

Around the year 2000 there were only a few display standards like PAL (601), NTSC (501), later HD (REC.709) and computer displays (sRGB). Display Referred also means that you see an “interpretation” of the scene that was shot in a studio or on location. And whatever the recording format was, in the Telecine step the material got initially squeezed into the dynamic range and color gamut of the resulting display – it gets “limited” in the first working steps.(Resolution comparison example: Think of a 8K camera, and the first step is a downscale to HD “1/4th” – you can’t get back to the full 8K resolution later on again.)

Working Scene Referred has a different approach. To see what the two methods have for an effect on an image, the change of a simple exposure operation is shown next. (Note: an exposure operation in Display Referred is non-linear whereas in Scene Referred it is linear)

Display Referred

A graded version of the image – one interpretation of the scene.Reducing the exposure by -1 stop. This means half the amount of light – this has a big effect in display referred but a rather small effect in scene referred.Reducing the exposure by -2 stops, this means only a quarter of the amount of light as the initial image. In display referred it already looks quite unnatural and dimmed.

Scene Referred

Scene Referred images are good to compute with, but not meant to be viewed like this.

ACES on the other hand has a different approach that is called “scene referred“. All material is uniformly “placed” in a much larger container in the file format Open-EXR.

This is true for the dynamic range and also for the color gamut.

In ACES you work with a “copy” of the scene that was shot, and not an “interpretation” of the scene. The quality of the “copy” depends on the ability of the camera model.„Scene referred“ means working with linear light values (actually, a camera sensor „counts“ photons).

One candle illuminates a scene with a certain brightness depending on the camera ISO/ASA settings, the aperture and shutter value. 10 or 100 candles illuminate the scene 10 or 100 times more. But because of how the eye and the human brain works, we don’t see the scene 10 or 100 times brighter.

Another example with modern LED light bulbs.

The left image is color balanced and exposure levelled image illuminated with only one LED bulb from a desk lamp.

The center and right are half as bright as the left left and look the same. Although they are made with two different bulbs. The right is the same as the left, but half the exposure time, whereas the center has the same exposure as the left, but a nearly half as bright bulb was used instead.

Default exposure in ACES – middle grey is at 18%. Left and right LED bulb have 1521 Lumen. Center LED bulb have 806 Lumen and is nearly half as bright as the other.The reference image: 18% grey has the value of 0.18.Half the exposure time results in half the values: 0.18 becomes 0.09.Same as the reference image, but with a bulb nearly half as bright(1521 lm / 806 lm ≈1.89 or factor of ≈0.53). 0.18 becomes 0.10.

ACES is viewed through a general tone-mapping method called RRT. Together with the display depended ODT you see the image on the screen. The ODT limits the material to a display specific dynamic range and color gamut. The underlying material has the full dynamic range (about 30 stops) and color gamut (every human visible color). This is more than any camera can record and any device can display.

The development of ACES has always in mind to make itself independent from any camera format and any display type. ACES can store images from the „perfect camera“ and show these images on the „perfect display“. Both of them don’t exist.

Therefore ACES is ready to create deliveries for different output formats like HD(REC.709), UHD Wide Gamut (REC.2020), Cinema DCI-P3, both in SDR (standard dynamic range) and HDR (high dynamic range). These formats you can find today on Netflix, the iTunes store and Amazon Prime Video for example.

Working in ACES with Maya is rather simple. The working color space is called ACEScg. The Arnold render viewer uses REC.709. Textures are converted to ACEScg before rendering. 3D software is working anyway in linear light. The rendered files are stored in the format Open-EXR in the colorspace ACEScg.

Compositing:

Nuke is set to work in ACES and the working color space is also ACEScg. The Nuke viewer is set to REC.709. Read nodes have to be set to the right color-space settings and a color matrix is applied to convert the images to ACEScg. Nuke is working since its creation in linear light long before a lot of other tools. The rendered files are stored in the format Open-EXR in the ACES storage format ACES 2065-1.

Grading:

Resolve can be also set to work in ACES.The working space is set to ACEScct. The ODT is set to REC.709.The graded files are exported as Open-EXR in the ACES storage format ACES 2065-1.

3D, Compositing and Grading are using the same file formats are all tools use the same View-Transform. This assures that everyone sees the same images and work with the same full quality files.

New possibilities and new workflows

Example: What if you need to deliver for a cinema project instead of TV/Web?

ACES pipeline switch from Rec.709 to cinema DCI-P3 example:

In ACES it is rather easy to switch to a full DCI-P3 cinema pipeline from REC.709. 3D & Compositing need new monitors that support the DCI-P3 color gamut. Grading would use a DCI-P3 projector to view the images. All tools set the ODT to DCI-P3. This is basically all what you need to do to switch from HD REC.709 to DCI-P3 cinema in an ACES color managed pipeline.

It’s that simple, because the underlying image data is untouched. It incompasses all available display types and might only need a little color-trim pass to get even more out of the actual display type.

A photographer/DOP and also a colourist can work best with all the elements in a picture. A finished well composed image should be the base to do a good grade.

In VFX you add green screen elements, add in 3D rendered elements or remove unwanted elements from a picture. This creates a new image. And this should be graded and not the empty plate.

The old saying shoot in camera what you can instead of adding stuff in post production is valid as always. By working scene referred we can add elements virtually on location instead of adding them onto or into an image on a “screen”.

This image represents an empty plate photography – ungradedA warm color grade like on an early morning. Without colorful elements this is fine. The compositing task: Adding two people, one of them with red shorts.This grade is now too oversaturated especially for the red trousers. The red is clipping!A little overall color grading adjustment with a lowering of the saturation and brightness of the whole image helps to make look the trousers and people better.

I hope this shows that a finished retouched and comped image with or without 3D rendered elements can differ so much from the original plate that you would judge and interpret the original plate and the compositing – to an extent that the graded result should look different.