LogoLogo
Reality 4.27
Reality 4.27
  • Reality 4.27
    • What's New
    • Introduction to Reality
    • Reality Walkthrough
      • Reality Examples Project
      • Tracking Calibration and Fine Tuning
      • Real-Time Ray Tracing
      • Raw Input
      • Augmented Reality Virtual Studio
      • Green Screen Virtual Studio
      • Setting up Realtime Video I/O
      • Projection Cube
      • Cyclorama Setup
      • Reality Keyer and Fine Tuning
        • Keyer FOV Softness
    • Reality Editor Guide
      • How Reality Editor Works?
      • Exporting UE4 Actors to RealityHub API
        • ZDActor Component
        • Exporting Blueprint Functions
        • Texture Pins on ZDActors
        • Exporting Properties
          • Directly Readable Variable Types
          • Property Update Callbacks
            • OnChanged_Event Callback
            • OnLengthChanged_Event Callback
            • OnChanged_ Function Callback
            • OnLengthChanged_ Function Callback
      • Exporting Assets to Reality API
        • Exporting Actors
        • Exporting Custom Blueprint Node
        • Exporting Custom Material Node
        • Exporting Lens Calibration
        • Exporting Post Process Material
        • Exporting Projection Material
        • Exporting Static Mesh
        • Exporting Unreal Motion Graphics
        • Exporting Unreal Texture Assets
      • Reality Custom Material Nodes
        • Defining the Variables in Material Definition
        • Custom Material Node
      • Using UMG in Reality
        • How to Use UMG with Reality
        • UMG Animations
        • Texture Brush
        • Asset Brush
        • Exporting Custom Events in UMG to Reality API
      • Additional ZD Components
        • ZDPieChart
        • ZDText
      • Helpful Links
      • Reality Editor Settings
        • Enabling DX12 and Ray Tracing
        • Hardware Accelerated Video Decoding
        • Enabling HAP Codec
    • Reality Setup Application
      • Audio
        • Audio Delay Node
        • Virtual Set Audio
        • Audio Mixer
      • Bypass Post Process
      • Fill and Key
        • Fill and Key using Dual-link
        • Fill and Key using Quad-link
      • Crop Function
      • Level Sequencer
      • Color Matrix Node
      • Media Playback
      • Preset for H.264 in Adobe Media Encoder
      • HAP Video Codec
      • Flow Control
      • GPUDirect Technology
    • Installation and Configuration Guide
      • Hardware Platforms
        • Legacy HP Z4G4 Configuration
          • Fan and Front Card Kit
          • Certified Power Supply
          • Video I/O Card Installation
            • Mellanox ConnectX®-5 Card Installation
            • AJA Video I/O Card Installation
          • GPU installation
            • Certified GPUs
            • Connecting GPU to 1000 watts power supply
              • GPU RTX 6000 installation (8+6 pin)
            • Connecting GPU to 750 watts power supply
              • GPU RTX 6000 installation (8+6 pin) Needs 6 to 8 Pin Adapter
          • Installing RAM
          • HP Performance Advisor
      • Adding Reality User to Windows
      • Downloading Distributions
      • Installing Steps
        • Upgrading
        • Installation
      • Configuration Steps
        • NVIDIA Configuration
        • AJA Video Card Configuration
        • R Drive Mapping
        • Internet Options
      • Updating RealityEngine AMPERE
    • Licensing Guide
      • Hardware Licensing
    • Reality Tutorials and Examples
      • Reality Editor Tutorials
        • Open/Close Door
        • Exporting UE Actor to ZD API (a.k.a ZDActor)
        • Custom Blueprint Node Tutorials
          • Using Event Tick Functions and Properties
          • Adding Custom Blueprint Nodes
        • Redirection of the Reality Engine Content
        • Creating Dynamic Materials
        • Changing ZDText with Dynamic String
        • Activating Destructible Mesh
        • Reality Texture Input
        • Activating Fire During Runtime
        • Asset Migration to Another Project
        • Rotating Door by Using Function Parameters
        • Creating Custom Cyclorama
      • Nodegraph Tutorials
        • Portal Window Setup
          • Adding Clipping Plane
          • How to Position Videowall
          • Setting Portal Window and Calibration
          • Creating the Augmented Part
    • Appendices
      • A - Camera Tracking and Lens Calibration
        • Camera Tracking Information
        • Tracking Device, Physical Connections
        • Lens Calibration
      • B - Standard Unreal Engine Post Processing Pipeline
      • C - Reality Keyer
        • What is Image Based Keying?
        • Reality Keyer Pipeline
        • Physical Factors Affecting the Key
      • D - Viewing Videomask
      • E - CLASS setting for Broadcast Cameras
      • F - ZD Blueprint Node Pin Reference
      • G - Utilizing TRAXIS talentS FreeD Data
    • Knowledge Base
    • Known Issues and Limitations
Powered by GitBook
On this page
  1. Reality 4.27
  2. Appendices

B - Standard Unreal Engine Post Processing Pipeline

PreviousLens CalibrationNextC - Reality Keyer

Last updated 2 years ago

Unreal Engine has a sophisticated post processing pipeline. Since this pipeline is very complex, it has its own node based processing system to do these processes. It is regenerated every frame, according to rendering parameters. For example, at one frame the bloom might be turned off (Bloom intensity set to 0), the other frame it might be turned on. Unreal Engine prepares its node-graph every frame, in order to optimize its processing speed.

“Translucency pass” is composited on top of “Scene Color pass” at the very early steps. Afterwards the depth of field is processed and “Separate translucency” is merged later on top of this image. This is why materials using “Separate translucency” are not affected by depth of field effect while standard translucent mate-rials are. Thus, separate translucent materials don’t look photo-realistic, since they do not match to the physical camera focus.

Separate translucency is also sometimes named as transparency after DOF. Using this type of translucent materials in a Virtual Studio environment will look fake, when used with Depth of Field.

Temporal anti-aliasing is applied after depth of field. “Temporal AA” is a very special anti-aliasing technique, which is very efficient in utilizing GPU hardware. We strongly suggest that you to use this anti-aliasing method for best quality results.

If the lens flares and/or bloom is effective in your post process settings, the bloom and lens flare passes will be calculated using anti-aliased image buffer. Bloom and lens flare buffers are merged and delivered into “Tone Mapper” pass. This pass first adds bloom / lens flare buffer on top of anti-aliased image, then tone maps the image from “linear gamma” space to “sRGB” color space.

Reality gives the option to output 16-bit HDR images from Unreal Engine’s “Tone Mapper”, while giving the option to bypass it. This gives a linear gamma output from Unreal Engine and gives higher bit depths for video output.

Reality can input “logarithmic encoded HDR” SDI video and convert it to “linear gamma” floating point HDR. Do all the processing in floating point precision and output back to SDI using logarithmic encoding. One example would be using “Arri Alexa/Amira” camera’s “LogC” encoding.