Da Vinci Resolve 15 New Features Guide
User Manual: Pdf
Open the PDF directly: View PDF .
Page Count: 282
Download | |
Open PDF In Browser | View PDF |
New Features Guide DaVinci Resolve 15 PUBLIC BETA Welcome Welcome to DaVinci Resolve 15 for Mac, Linux and Windows! DaVinci is the world’s most trusted name in color and has been used to grade more Hollywood films, TV shows, and commercials than anything else. Now, with DaVinci Resolve 15, you get a complete set of editing, advanced color correction, professional Fairlight audio post production tools and now Fusion visual effects combined in one application so you can composite, edit, grade, mix and master deliverables from start to finish, all in a single tool! DaVinci Resolve 15 has the features professional editors, colorists, audio engineers and VFX artists need, and is built on completely modern technology with advanced audio, color and image processing that goes far beyond what any other system can do. With this release, we hope to inspire creativity by letting you work in a comfortable, familiar way, while also giving you an entirely new creative toolset that will help you cut and finish projects at higher quality than ever before! We hope you enjoy reading this manual. With its customizable interface and keyboard shortcuts, DaVinci Resolve 15 is easy to learn, especially if you’re switching from another editor, and has all of the tools you need to create breathtaking, high end work! The DaVinci Resolve Engineering Team Grant Petty CEO Blackmagic Design About This Guide DaVinci Resolve 15 is a huge release, with the headlining addition of the Fusion page for advanced effects and motion graphics, and workflow‑enhancing features added in equal measure to the Edit page, the Color page, and the Fairlight page. Consequently, this year’s public beta needed a different approach to the documentation. This “New Features Guide” provides a focused and comprehensive look at only the new features in DaVinci Resolve 15. If you’re a Resolve user already and you just want to know what’s changed, this is for you. If you’re a new user who needs to learn the fundamentals, you should refer to the DaVinci Resolve 14 Manual, but know that none of the new features in version 15 are covered there. This PDF is divided into two parts. Part 1 covers new features in the previously existing pages of DaVinci Resolve, including chapters on overall interface enhancements, Edit page improvements, Color page features, and Fairlight page additions. Part 2 provides a “documentation preview” of chapters that cover the basics of the new Fusion page. Fusion compositing is a huge topic, and the chapters of Part 2 seek to provide guidance on the fundamentals of working in Fusion so that you can try these features for yourself, even if you don’t have a background in node-based compositing. Eventually, the information in this guide will be rolled into the next update of the manual that will accompany the final release of DaVinci Resolve 15. For now, use this guide as a tour of the exciting new features being unveiled, and have fun. DaVinci Resolve 15 BETA New Features Guide 3 Contents PART 1 New Features in 15 1 General Improvements 1-2 2 Edit Page Improvements 1-15 3 Subtitles and Closed Captioning 1-37 4 Color Page Improvements 1-49 5 New ResolveFX 1-68 6 Fairlight Page Improvements 1-81 7 FairlightFX 1-105 PART 2 Fusion Page Manual Preview 8 Introduction to Compositing in Fusion 2-2 9 Using the Fusion Page 2-7 10 Getting Clips into the Fusion Page 2-46 11 2-52 Image Processing and Color Management 12 Understanding Image Channels and Node Processing 2-58 13 Learning to Work in the Fusion Page 2-84 14 Creating Fusion Templates 2-150 DaVinci Resolve 15 BETA New Features Guide 4 PART 1 New Features in 15 Chapter 1 General Improvements Part I of the DaVinci Resolve 15 New Features Guide explores and explains all of the new features that have been added to the Media, Edit, Color, Fairlight, and Deliver pages in this year’s public beta. This particular chapter covers overall enhancements that affect the entire application. DaVinci Resolve 15 introduces numerous improvements to saving, to the overall user interface, to Project Settings and Preferences, and to export. Furthermore, many improvements to performance have been implemented to make working in DaVinci Resolve even faster. There are a variety of image quality enhancements, including support for “Super Scale” image enlargement for doing higher-quality enlargements when you’re dealing with smaller-resolution archival footage or when you’re punching way into a clip to create an emergency closeup. There’s also support for numerous additional media formats, with support for built-in IMF encoding and decoding, and the beginning of workflow‑oriented scripting support for DaVinci Resolve. Chapter – 1 General Improvements 1-2 Contents Overall User Interface Enhancements 1-4 Menu Bar Reorganization 1-4 Contextual Menu Consolidation 1-4 Page-Specific Keyboard Mapping 1-4 Project Versioning Snapshots 1-5 Ability to Open DRP Files From the macOS Finder 1-6 Floating Timecode Window 1-6 Ability to Minimize Interface Toolbars and the Resolve Page Bar 1-7 Performance Enhancements 1-7 Selective Timeline and Incremental Project Loading 1-7 Bypass All Grades Command Available on All Pages 1-8 Improved PostgreSQL Database Optimization 1-8 Optimized Viewer Updates 1-8 Improved Playback on Single GPU Systems Showing Scopes 1-8 Improved Playback With Mismatched Output Resolution and Video Format 1-8 Audio I/O Processing Block Size 1-8 Support for OpenGL Compute I/O on Supported Systems 1-9 Video Stabilization Has Been GPU Accelerated 1-9 ResolveFX Match Move is GPU Accelerated 1-9 Quality Enhancements 1-9 “Super Scale” High Quality Upscaling 1-9 Improved Motion Estimation for Retime and Noise Reduction Effects 1-10 Audio I/O Enhancements 1-11 Full Fairlight Engine Support For the Edit Page 1-11 Support for Native Audio on Linux 1-11 Record Monitoring Using the Native Audio Engine 1-11 Media and Export Improvements 1-11 Improved Media Management for Temporally Compressed Codecs 1-11 Support for Frame Rates Up to 32,000 Frames Per Second 1-11 Support for XAVC-Intra OP1A HDR Metadata 1-12 Support for ARRI LF Camera Files 1-12 Support for HEIC Still Image Media 1-12 Support for TGA Files 1-12 Support for DNX Metadata in QuickTime Media 1-12 Kakadu-based JPEG2000 Encoding and Decoding (Studio Only) 1-12 Native IMF Encoding and Decoding (Studio Only) 1-13 Native Unencoded DCP Encoding and Decoding (Studio Only) 1-14 Scripting Support for DaVinci Resolve 1-14 Chapter – 1 General Improvements 1-3 Overall User Interface Enhancements A number of usability enhancements improve command access and keyboard shortcuts throughout every page of DaVinci Resolve, as well as facilitate the backing up of your projects in DaVinci Resolve. Menu Bar Reorganization In an effort to accommodate the additional functionality of the Fusion and Fairlight pages with their attendant menus, all commands from the Nodes menu have been moved into the Color menu, making room for the new Fusion menu on laptops with limited screen real estate. The menus have been updated to accommodate the Fusion and Fairlight pages Additionally, most menus have been reorganized internally to place multiple variations on commands within submenus, making each menu less cluttered so that commands are easier to spot. Contextual Menu Consolidation Contextual menus throughout DaVinci Resolve have been consolidated to omit commands that were formerly disabled, with the result being shorter contextual menus showing only commands that are specific to the area or items you’ve right-clicked on. Page-Specific Keyboard Mapping When customizing keyboard shortcuts, you can now specify whether a keyboard shortcut is “Global” so that it works identically on every page, or you can map a particular keyboard shortcut to do a particular thing on a specific page. With page-specific keyboard shortcuts, you can have a single key do different things on the Edit, Fusion, Color, or Fairlight pages. Keyboard shortcuts can now be mapped to specific pages, if you like Chapter – 1 General Improvements 1-4 Project Versioning Snapshots Turning on the Project Backups checkbox in the Project Save and Load panel of the User Preferences enables DaVinci Resolve to save multiple backup project files at set intervals, using a method that’s analogous to a GFS (grandfather father son) backup scheme. This can be done even while Live Save is turned on. Each project backup is a complete project file, excluding stills and LUTs. The User Preferences controls for Project Backups Project backups are only saved when changes have been made to a project. If DaVinci Resolve sits idle for any period of time, such as when your smart watch tells you to go outside and walk around the block, no additional project backups are saved, preventing DaVinci Resolve from overwriting useful backups with unnecessary ones. Three fields let you specify how often to save a new project backup. Perform backups every X minutes: The first field specifies how often to save a new backup within the last hour you’ve worked. By default, a new backup is saved every 10 minutes, resulting in six backups within the last hour. Once an hour of working has passed, an hourly backup is saved and the per-minute backups begin to be discarded on a “first in, first out” basis.” By default, this means that you’ll only ever have six backups at a time that represent the last hour’s worth of work. Hourly backups for the past X hours: The second field specifies how many hourly project backups you want to save. By default, 8 hourly backups will be saved for the current day you’re working, which assumes you’re working an eight hour day (wouldn’t that be nice). Past that number, hourly backups will begin to be discarded on a “first in, first out” basis. Daily backups for the past X days: The third field specifies for how many days you want to save backups. The very last project backup saved on any given day is preserved as the daily backup for that day, and by default daily backups are only saved for five days (these are not necessarily consecutive if you take some days off of editing for part of the week). Past that number, daily backups will begin to be discarded on a “first in, first out” basis. If you’re working on a project over a longer stretch of time, you can always raise this number. Project backup location: Click the Browse button to choose a location for these project backups to be saved. By default they’re saved to a “ProjectBackup” directory on your scratch disk. Chapter – 1 General Improvements 1-5 Once you’ve enabled Project Backups for a long enough time, saved project backups are retrievable in the Project Manager, via the contextual menu that appears when you right-click a project. Opening a project backup does not overwrite the original project; project backups are always opened as independent projects. Restoring a project backup in the Project Browser NOTE: When using this feature, the very first project backup that’s saved for a given day may be a bit slow, but all subsequent backups should be unnoticeable. Ability to Open DRP Files From the macOS Finder This is a feature that’s specific to macOS. If you double-click a DaVinci Resolve .drp file in the Finder, this will automatically open DaVinci Resolve, import that project into the File Browser, and open that project so that you’re ready to work. Floating Timecode Window A Timecode Window is available from the Workspace menu on every page. Choosing this option displays a floating timecode window that shows the timecode of the Viewer or Timeline that currently has focus. This window is resizable so you can make the timecode larger or smaller. A new floating timecode window is available Chapter – 1 General Improvements 1-6 Ability to Minimize Interface Toolbars and the Resolve Page Bar If you right-click anywhere within the UI toolbar at the top of each page, or the Resolve Page Bar at the bottom of the DaVinci Resolve UI, you “Show Icons and Labels” or “ Show Icons Only.” If you show icons only, the Resolve Page Bar at the bottom takes less room, and the UI toolbar becomes less cluttered. The UI Toolbar and Page bar showing icons only, to save space Performance Enhancements Several new features have been added to improve performance. Selective Timeline and Incremental Project Loading To improve the performance of longer projects with multiple timelines, the “Load all timelines when opening projects” checkbox in the Project Save and Load panel of the User Preferences defaults to off. When this checkbox is off, opening a project only results in the last timeline you worked on being opened into memory; all other timelines are not loaded into RAM. This speeds up the opening of large projects. However, you may experience brief pauses when you open other timelines within that project, as each new timeline must be loaded into RAM as you open it. If you open a particularly gigantic timeline, a progress bar will appear letting you know how long it will take to load. Another advantage of this is the reduction of each project’s memory footprint, which is particularly valuable when working among multiple projects using Dynamic Project Switching. If you turn this on, all timelines will be loaded into RAM and you’ll experience no pauses when opening timelines you haven’t opened already. However, projects with many timelines may take longer to open and save. The new “Load all timelines when opening projects” preference, that defaults to off to speed up project load times NOTE: While “Load all timelines when opening projects” is turned on, the Smart and User Cache become unavailable. Chapter – 1 General Improvements 1-7 Bypass All Grades Command Available on All Pages The Bypass All Grades command, previously available only on the Color page, is now available on the Edit page either via View > Bypass All Grades, or via a button in the Timeline Viewer. Turning off grades is an easy way to improve performance on low power systems, and it’s also a convenient way to quickly evaluate the original source media. The Bypass All Grades button in the Timeline Viewer title bar of the Edit page Improved PostgreSQL Database Optimization The Optimize command for PostgreSQL project databases in the Database sidebar of the Project window produces improved results for large projects, resulting in better project load and save performance. Optimized Viewer Updates A new preference, “Optimized Viewer Updates,” which only appears on multi-GPU macOS and Windows systems, and on single- or multi-GPU Linux systems, enables faster viewer update performance. Improved Playback on Single GPU Systems Showing Scopes Video scope playback performance is improved on single-GPU systems. Improved Playback With Mismatched Output Resolution and Video Format Playback performance has been greatly improved when you set Output Resolution in the Image Scaling panel of the Projects settings to a different frame size than Video Format in the Master Settings panel, while outputting to any Blackmagic Design capture and playback device. Audio I/O Processing Block Size A new “Audio Processing Block Size X Samples” parameter in the Hardware Configuration panel of the System Preferences lets you increase the sample block size to add processing headroom to the system, at the expense of adding latency to audio playback. The default value is Auto, which automatically chooses a suitable setting for the audio I/O device you’re using. For those who have specific needs and are interested in setting this manually, here are some examples of use. In a first example, when a system is under a heavy load (there are many plug-ins being used on many tracks), then increasing the block size to add processing headroom will result in a longer delay every time your audio hardware requests samples to feed the speakers. If you’re only mixing, the resultant latency may not be a problem, so this gives you the option to add headroom so your system can run a few more plug-ins or tracks. Chapter – 1 General Improvements 1-8 On the other hand, this increased delay in the processed audio running through the mixer is a much bigger problem if you’re recording an artist in an ADR session, where they need to hear themselves in the headphones, or when you’re recording foley or voice over and there’s an increased delay between what you see and what you’re recording, so in this case sticking with the default value (or smaller) will sacrifice processing headroom for diminished latency. TIP: A common strategy when you need to force more cooperation from a particular workstation and audio interface is to reduce Audio Processing Block Size when you’re about to do a recording session, when track and plug-in use is lower. Later, when you start mixing in earnest and adding plug-ins, you can increase Audio Processing Block Size to give you better performance once you’re finished recording. Support for OpenGL Compute I/O on Supported Systems The “GPU processing mode” pop-up menu in the Hardware Configuration panel of the System Preferences has a new option for OpenCL on supported systems. Video Stabilization Has Been GPU Accelerated Video stabilization in DaVinci Resolve has been accelerated, demonstrating an up to 6x improvement vs. 14.3 on the Late 2014 Retina 5K iMac. Additionally, the automatic cropping behavior has been improved. ResolveFX Match Move is GPU Accelerated The Match Move ResolveFX plug-in is now GPU Accelerated, resulting in faster tracking and compositing workflows on the Color page. Quality Enhancements Two image quality enhancements have been incorporated providing higher visual quality for some of the most challenging operations in DaVinci Resolve: upscaling, retiming, and noise reduction. “Super Scale” High Quality Upscaling For instances when you need higher-quality upscaling than the standard Resize Filters allow, you can now enable one of three “Super Scale” options in the Video panel of the Clip Attributes window for one or more selected clips. The Super Scale pop-up menu provides three options of 2x, 3x, and 4x, as well as Sharpness and Noise Reduction options to tune the quality of the scaled result. Note that all of the Super Scale parameters are in fixed increments; you cannot apply Super Scale in variable amounts. Super Scale options in the Video panel of the Clip Attributes Chapter – 1 General Improvements 1-9 Selecting one of these options enables DaVinci Resolve to use advanced algorithms to improve the appearance of image detail when enlarging clips by a significant amount, such as when editing SD archival media into a UHD timeline, or when you find it necessary to enlarge a clip past its native resolution in order to create a closeup. This is an extremely processor-intensive operation, so be aware that turning this on will likely prevent real-time playback. One way to get around this is to create a string-out of all of the source media you’ll need to enlarge at high-quality, turn on Super Scale for all of them, and then render that timeline as individual clips, while turning on the “Render at source resolution” and “Filename uses > Source Name” options. Improved Motion Estimation for Retime and Noise Reduction Effects When using mixed frame rate clips in a timeline that has Optical Flow retiming enabled, when using Optical Flow to process speed change effects, or when using Image Stabilization or Temporal Noise Reduction controls in the Color page, the Motion Estimation pop-up of the Master Settings (in the Project Settings window) lets you choose options that control the trade-off between speed and quality. There are additional “Enhanced” Optical Flow settings available in the “Motion estimation mode” pop-up in the Master Settings panel of the Project Settings. The “Standard Faster” and “Standard Better” settings are the same options that have been available in previous versions of DaVinci Resolve. They’re more processor-efficient and yield good quality that are suitable for most situations. However, “Enhanced Faster” and “Enhanced Better” should yield superior results in nearly every case where the standard options exhibit artifacts, at the expense of being more computationally intensive, and thus slower on most systems. New improved Motion estimation mode settings in the Master Settings panel of the Project Settings Chapter – 1 General Improvements 1-10 Audio I/O Enhancements Audio processing, playback, and recording in DaVinci Resolve 15 has been improved for better cross-platform support. Full Fairlight Engine Support For the Edit Page All Edit page features now use the Fairlight audio engine, providing superior transport control performance, as well as the ability to choose which audio I/O device to output to. Support for Native Audio on Linux DaVinci Resolve on Linux now supports using your workstation’s system audio on Linux, instead of only supporting Decklink audio, as with previous systems. This means that DaVinci Resolve can use your Linux workstation’s on-board audio, or any Advanced Linux Sound Architecture (ALSA)-supported third party audio interface. Record Monitoring Using the Native Audio Engine The native audio of your workstation’s operating system can now be used for record monitoring. This makes it possible to set up recording sessions where your audio input is patched via a third party interface, and the audio output you’re monitoring can be patched via your computer’s native audio. Media and Export Improvements DaVinci Resolve 15 supports several new media formats and metadata encoding methods. Improved Media Management for Temporally Compressed Codecs As of DaVinci Resolve 15, the Media Management window can “copy and trim” clips using temporally compressed media codecs such as H.264, XAVC, and AVC-Intra, enabling you to eliminate unused media for these formats during media management without recompressing or transcoding. Support for Frame Rates Up to 32,000 Frames Per Second To accommodate media from different capture devices capable of high-frame-rate slow motion capture, and to future-proof DaVinci Resolve against the rapidly improving array of capture devices being developed every year, DaVinci Resolve has shifted the upper limit of supported frame rates to 32,000 frames per second. Hopefully that’ll cover things for a while. Please note, just because extremely high frame rate media is supported, do not expect real time performance at excessively high frame rates, and understand that what performance your workstation is capable of depends on its configuration and the speed of your storage. Chapter – 1 General Improvements 1-11 Support for XAVC-Intra OP1A HDR Metadata DaVinci Resolve 15 now supports writing color space and gamma metadata to MXF OP1A format media using the XAVC-Intra codecs. Support for ARRI LF Camera Files Media from the new ARRI LF (Large Format) camera is now supported at all resolutions and frame rates. Support for HEIC Still Image Media DaVinci Resolve 15 supports the High Efficiency Video Compression (HEIF) standard used by Apple for capturing images on newer iPhones. With HEIF support, these still images can be used in DaVinci Resolve projects. Support for TGA Files TGA stills and image sequences are supported for import. Support for DNX Metadata in QuickTime Media DaVinci Resolve 15 supports DNX Metadata including Color Range, Color Volume, and HasAlpha within QuickTime files. Kakadu-based JPEG2000 Encoding and Decoding (Studio Only) DaVinci Resolve 15 supports the encoding and decoding of JPEG2000 using a library licensed from Kakadu software. This includes a complete implementation of the JPEG2000 Part 1 standard, as well as much of Parts 2 and 3. JPEG2000 is commonly used for IMF and DCP workflows. A variety of Kakadu JPEG2000 options are available when you choose MJ2, IMF, JPEG 2000, MXF OP-Atom, MXF OP1A, or QuickTime from the Format pop-up menu of the Video panel of the Render Settings on the Deliver page. Choosing a JPEG2000-compatible format enables Kakadu JPEG 2000 codec options Chapter – 1 General Improvements 1-12 Native IMF Encoding and Decoding (Studio Only) The Format pop-up in the Video panel of the Render Settings now has a native IMF option that lets you export to the SMPTE ST.2067 Interoperable Master Format (IMF) for tapeless deliverables to networks and distributors. No additional licenses or plug-ins are required to output to IMF. The IMF format supports multiple tracks of video, multiple tracks of audio, and multiple subtitle and closed caption tracks, all of which are meant to accommodate multiple output formats and languages from a single deliverable. This is done by wrapping a timeline’s different video and audio tracks (media essences) and subtitle tracks (data essences) into a “composition” within the Material eXchange Format (MXF). When the IMF format is selected, the Codec pop-up menu presents numerous options for Kakadu JPEG2000 output, including RGB, YUV, and Dolby Vision options. Additional controls appear for Maximum bit rate, Lossless compression, the Composition name, the Package Type (current options include App2e), and a Package ID field. Render settings in the Export Video section for the IMF format Chapter – 1 General Improvements 1-13 Native Unencoded DCP Encoding and Decoding (Studio Only) DaVinci Resolve 15 also has new native DCP Encoding and Decoding support built-in, for unencoded DCP files only. That means that you can output and import (for test playback) unencoded DCP files without needing to purchase a license of EasyDCP. If you have a license, a preference enables you to choose whether to use EasyDCP (for creating encrypted DCP output), or the native Resolve encoding. Native DCP settings in Resolve Scripting Support for DaVinci Resolve As of DaVinci Resolve 15, Resolve Studio begins to add support for Python and LUA scripting to support various kinds of workflow automation. More information will be forthcoming as it becomes available via developer documentation. Chapter – 1 General Improvements 1-14 Chapter 2 Edit Page Improvements The Edit page sees a wide variety of improvements and enhancements to nearly every area of editing in DaVinci Resolve 15. Media Pool improvements include new editable Clip names by default and larger thumbnails. Marker enhancements include drawn annotations and the ability to turn markers with duration into In and Out points. Editing enhancements are far-reaching, beginning with the ability to stack multiple timelines to accommodate so-called “pancake editing” techniques, improved dynamic trim behaviors with automatic edit selection and the ability to move to next and previous edits, the ability to create subclips via drag and drop from the Source Viewer to the Media Pool, and many, many more improvements large and small. Finally, editorial effects enhancements include the new TextPlus generator, keyframable OpenFX and ResolveFX, alpha transparency support in compound clips, and improved optical flow options for speed effects. Chapter – 2 Edit Page Improvements 1-15 Contents Media Pool and Clip Display Enhancements 1-18 Display Name is Now Called Clip Name, On by Default 1-18 Display Audio Clip Waveforms in Media Pool and Media Storage 1-18 Media Pool Command for Finding Synced Audio Files 1-19 Improved Media Pool Column Customization 1-19 Larger Thumbnails in the Media Pool 1-19 Recent Clips Pop-Up Menu 1-19 Ability to Create Subclips via Drag and Drop from Source Viewer 1-20 Ability to Remove Subclip Limits 1-20 Import Hierarchically Organized Nests of Empty Directories 1-20 Import Clips with Metadata via Final Cut Pro 7 XML 1-20 Enhancements to Markers and Flags 1-21 Drawn Annotations on the Viewer 1-21 Ability to Create Markers and Flags With Specific Colors 1-22 Timecode Entry in Marker Dialogs 1-22 Clip Markers Show Overlays in the Timeline Viewer 1-22 Command to Turn Markers With Duration into In and Out Points 1-23 Editing Enhancements 1-23 Tabbed and Stacked Timelines 1-23 Improved Dynamic Trim Behaviors 1-25 Ability to Modify Clip Duration Via Timecode 1-26 Ability to Delete Multiple Timeline Gaps at Once 1-26 Improved Separation Between the Video and Audio Tracks 1-27 Improved Ripple Cut and Ripple Delete Behavior 1-27 Improved Automatic Audio Track Creation 1-28 New Play Again Command 1-28 Option to “Stop and Go To Last Position” 1-28 Single Viewer Mode is Available in Dual Screen Layout 1-28 Copy and Paste Timecode in Viewer Timecode Fields 1-28 Improved Timecode Entry 1-29 Replace Edit Using Clips Already in the Timeline 1-29 Option to Ripple the Timeline in Paste Attributes 1-30 Transition Categories in the Effects Library 1-30 Linked Move Across Tracks 1-30 Track Destination Keyboard Shortcuts Also Disable Tracks 1-31 Ability to Mark In and Out Points in Cinema Mode 1-31 Updated Timeline View Options Menu in the Toolbar 1-31 Chapter – 2 Edit Page Improvements 1-16 Edit Page Effects Enhancements 1-32 The Text+ Title Generator in the Edit Page 1-32 Fusion Titles and Fusion Templates 1-34 Support for Caching of Titles and Generators 1-35 Position Curves in the Timeline Curve Editor 1-35 Keyframable OpenFX and ResolveFX 1-35 Compound Clips Support Alpha Transparency 1-36 Improved Smooth Cut 1-36 Improved Optical Flow for Speed Effects 1-36 Chapter – 2 Edit Page Improvements 1-17 Media Pool and Clip Display Enhancements A variety of improvements have been made to clip display and the Media Pool. Display Name is Now Called Clip Name, On by Default The clip metadata that was formerly called “Display Name” is now known as “Clip Name,” so that there are two sets of clip identification metadata available to DaVinci Resolve, Clip Name and File Name, both of which are visible in the Media Pool in List View. Starting in DaVinci Resolve 15, Clip Name is the default clip identifier, while File Name is hidden by default. This means that you can freely edit the default name that’s visible in the Media Pool, either in List or Icon view, without needing to change modes. When necessary, you can switch DaVinci Resolve to identify clips by file name only by choosing View > Show File Names. TIP: Don’t forget that you can use % variables to automatically populate Clip Names via metadata, such as Scene, Take, and Description (%Scene_%Take_%Description), leveraging whatever metadata entry you’ve done to prepare your clips to automatically create useful clip naming conventions. Using Clip Attributes, you can edit the Clip Name for multiple selected clips using metadata variables all at once, which is a real time‑saver. Display Audio Clip Waveforms in Media Pool and Media Storage The Media Pool option-menu presents an option to Show Audio Waveforms. When you do so, every audio clip in the Media Pool appears with an audio waveform within its thumbnail area. If Live Media Preview is on in the Source Viewer, you can then scrub through each clip and hear its contents. If you don’t want to see audio waveforms, you can turn this option off. You can now enable waveform thumbnails in the Media Pool that you can scrub with Live Media Preview Chapter – 2 Edit Page Improvements 1-18 Media Pool Command for Finding Synced Audio Files When you’ve synced dual-system audio and video clips together in DaVinci Resolve, you can find the audio clip that a video clip has been synced to using the following procedure. To find the audio clip that a video clip has been synced to: Right-click a video clip that’s been synced to audio, and choose “Reveal synced audio in Media Pool” from the contextual menu. The bin holding the synced audio clip is opened and that clip is selected. Improved Media Pool Column Customization The list of Media Pool columns that appears when you right-click on one of the column headers is alphabetized, making it easier to find the columns you need when you’re choosing which columns to show or hide. Larger Thumbnails in the Media Pool The maximum size of thumbnails in the Media Pool has been increased. The maximum size of thumbnails has increased Recent Clips Pop-Up Menu A new pop-up at the top of the Source Viewer, next to the name of the currently open clip, lets you open a menu containing a list of the last 10 clips you opened in the Source Viewer. This list is first in, first out, with the most recently opened clips appearing at the top. A recent clips menu lets you recall the last 10 clips you opened in the Source Viewer Chapter – 2 Edit Page Improvements 1-19 Ability to Create Subclips via Drag and Drop from Source Viewer There’s a new way of creating subclips. Simply open a clip into the Source Viewer, set In and Out points, and then drag a clip from the Source Viewer to the Media Pool. The new clip that’s added to the Media Pool will be a subclip with a total duration that’s bounded by the duration you marked. Ability to Remove Subclip Limits You can right-click any subclip in the Media Pool and choose Edit Subclip to open a dialog in which you can turn on a checkbox to use the subclip’s full extents, or to change the start or end timecode of the subclip via timecode fields, before clicking Save to modify the subclip. The Edit Subclip dialog Import Hierarchically Organized Nests of Empty Directories You can import a nested series of directories and subdirectories that constitutes a default bin structure you’d like to bring into projects, even if those directories are empty, by dragging them from your file system into the Media Pool bin list of a project. The result is a hierarchically nested series of bins that mimics the structure of the directories you imported. This is useful if you want to use such a series of directories as a preset bin structure for new project. Import Clips with Metadata via Final Cut Pro 7 XML In order to support workflows with media asset management (MAM) systems, DaVinci Resolve supports two additional Media Pool import workflows that use Final Cut Pro 7 XML to import clips with metadata. To import clips with metadata using Final Cut Pro 7 XML files, do one of the following: Right-click anywhere in the background of the Media Pool, choose Import Media from XML, and then choose the XML file you want to use to guide import from the import dialog. Drag and drop any Final Cut Pro 7 XML file into the Media Pool from the macOS Finder. Every single clip referenced by that XML file that can be found via its file path will be imported into the Media Pool, along with any metadata entered for those clips. If the file path is invalid, you’ll be asked to navigate to the directory with the corresponding media. Chapter – 2 Edit Page Improvements 1-20 Enhancements to Markers and Flags The use of markers and flags to identify moments in clips and things to keep track of has been significantly enhanced in DaVinci Resolve 15. Drawn Annotations on the Viewer It’s now possible to use the Annotations mode of the Timeline Viewer to draw arrows and strokes of different weights and colors directly on the video frame, in order to point out or highlight things that need to be fixed. These annotations are stored within markers, similarly to marker names and notes. To start, simply choose Annotations mode from the Timeline Viewer mode pop-up menu to do this. Choosing Annotations from the Viewer Mode pop-up menu Once in Annotations mode, an Annotations toolbar appears showing the following options: Draw tool with line weight pop-up: Click the Draw tool to be able to free-form draw on the Viewer. Click the Line Weight pop-up to choose from one of three line weights to draw with. Arrow tool: Click the Arrow tool to draw straight-line arrows pointing at features you want to call attention to. Arrows are always drawn at the same weight, regardless of the weight selected for the Line tool. Color pop-up: Choose a color for drawing or lines. The annotations toolbar in the Viewer Methods of making and editing annotations: To create an annotation: Simply enable Annotations mode, then park the playhead on any frame of the Timeline and start drawing. A marker will automatically be added to the Timeline at that frame, and that marker contains the annotation data. If you park the playhead over a preexisting Timeline marker, annotations will be added to that marker. To edit a stroke or arrow you’ve already created: Move the pointer over a stroke or arrow and click to select it, then choose a new line weight or color from the appropriate pop-up menu, or drag that stroke or arrow to a new location to move it. To delete a stroke or arrow: Move the pointer over a stroke or arrow and click to select it, then press the Delete or Backspace keys. Chapter – 2 Edit Page Improvements 1-21 Drawing annotations to highlight feedback Ability to Create Markers and Flags With Specific Colors The Mark > Add Marker and Add Flag submenus have individual commands for adding markers and flags of specific colors directly to clips and the Timeline. Additionally, these individual commands can be assigned specific keyboard shortcuts if you want to be able to place a specific marker color or flag color at a keystroke. Individually mappable marker color commands Timecode Entry in Marker Dialogs The Time and Duration timecode fields of marker dialogs can now be manually edited, to numerically move a marker, or to create a marker with a specific duration. Furthermore, the timecode in these fields can be copied from or pasted to. Clip Markers Show Overlays in the Timeline Viewer Clip markers and Timeline markers both appear as Timeline Viewer overlays when Show Marker Overlays is enabled in the Timeline Viewer option menu. Pressing Shift-Up or Down Arrow moves the playhead back and forth among both Timeline and Clip markers in the currently open Timeline. Chapter – 2 Edit Page Improvements 1-22 Command to Turn Markers With Duration into In and Out Points Two commands, in the contextual menu of the Source Viewer Jog Bar, work together to let you turn In and Out points into Markers with Duration, and vice versa: Convert In and Out to Duration Marker: Turns a pair of In and Out points into a marker with duration. By default, no key shortcut is mapped to this command, but you can map one if you like. Convert Duration Marker to In and Out: Turns a marker with duration into a pair of In and Out points, while retaining the marker. By default, no key shortcut is mapped to this command, but you can map one if you like. Using these two commands, you can easily use markers with duration to mark regions of clips that you want to log for future use, turning each region into an In and Out point when necessary for editing. Editing Enhancements Quite a few enhancements improve the editing experience in the Edit page. Tabbed and Stacked Timelines The Timeline now supports the option to have tabs that let you browse multiple timelines quickly. With tabbed timeline browsing enabled, a second option lets you open up stacked timelines to simultaneously display two (or more) timelines one on top of another. Tabbed Timelines The Timeline View Options menu in the toolbar has a button that lets you enable tabbed browsing and the stacking of timelines. A button in the Timeline View Options enables tabbed timelines When you first turn this on, a Timeline tab bar appears above the Timeline, displaying a tab for the currently open timeline that contains a Close button and a Timeline pop-up menu. Once you enable Tab mode, opening another timeline from the Media Pool opens it into a new tab. Chapter – 2 Edit Page Improvements 1-23 To the right of the currently existing tabs, an Add Tab button lets you create additional tabs that default to “Select Timeline.” Click any tab’s pop-up menu to choose which timeline to display in that tab. Tabs above the timeline editor let you switch among multiple timelines quickly Methods of working with tabbed timelines: Click any tab to switch to that timeline. Use the pop-up menu within any tab to switch that tab to display another timeline from the Media Pool. Each tab’s pop-up menu shows all timelines within that project, in alphabetical order, but a timeline can only be open in one tab or stack at a time. Drag any tab left or right to rearrange the order of timeline tabs. Click any tab’s Close button to close that timeline and remove that tab. Stacked Timelines While tabbed browsing is turned on, an Add Timeline button appears to the right that enables you to stack two (or more) timelines one on top of another. This lets you have two (or more) timelines open at the same time, making it easy to edit clips from one timeline to another. A good example of when this is useful is when you’ve created a timeline that contains a stringout of selects from a particular interview. You can stack two Timeline Editors, one on top of another, and then open the Selects Timeline at the top and the Timeline you’re editing into at the bottom. With this arrangement, it’s easy to play through the top timeline to find clips you want to use, to drag and drop into the bottom timeline to edit into your program. Two timelines stacked on top of one another Chapter – 2 Edit Page Improvements 1-24 To enable or disable stacked timelines: Click the Add Timeline button at the right of the Timeline tab bar. The button for adding a stacked timeline Once you’ve enabled stacked timelines, each timeline has its own tab bar and an orange underline shows which timeline is currently selected. At the right of each Timeline tab bar, a Close Timeline button appears next to the Add Timeline button, which lets you close any timeline and remove that timeline browsing area from the stack. The button for closing a stacked timeline Improved Dynamic Trim Behaviors The Dynamic Trim mode that lets you use JKL playback behaviors to do resize or ripple trimming to one or more selected edits has been improved in a number of key ways. Pressing W to Enable Dynamic Trimming Automatically Selects the Nearest Edit If no edit is currently selected in the Timeline, then pressing W to enable Dynamic Trim automatically selects the nearest edit, similarly to if you pressed V to initiate the Select Nearest Edit Point command. If you’ve already selected an edit, or made a multiple-edit-point selection, nothing changes and the edit points you’ve selected will be used for dynamic trimming. By default, the entire edit is selected, positioning you to perform a dynamic Roll edit. However, you can press the U key (Edit Point Type) to toggle the edit selection among three positions, in order to trim the outgoing half or incoming half of the selected edit point. Also by default, both the Video and Audio of the current edit are selected if Linked Selection is enabled. Pressing Option-U (Toggle V+A/V/A) lets you toggle the edit selection to include both Video+Audio, Video only, or Audio only. When you’re finished trimming, whatever edit point type and A/V toggle you used last is remembered by DaVinci Resolve, and will be selected the next time you enable Dynamic Trimming. You Can Use the Next and Previous Edit Commands While in Dynamic Trim Mode Previously, you were unable to use the Up and Down Arrow keys to move the selected edit from one edit point to another. This has been fixed so you can now move the selected edit forward and back in your timeline, using Dynamic Trim to adjust as many edits as you like. Chapter – 2 Edit Page Improvements 1-25 Ability to Modify Clip Duration Via Timecode You can change a clip’s duration numerically in one of two ways. To change a selected clip’s duration: 1 Decide if you want to ripple the Timeline or overwrite neighboring clips when you change a clip’s duration. If you want to ripple the Timeline, choose the Trim tool. If you want to overwrite neighboring clips or leave a gap, choose the Selection tool. 2 Do one of the following: Select a clip and choose Clip > Change Clip Duration Right-click any clip in the Timeline and choose Change Clip Duration from the contextual menu 3 When the Edit Duration Change dialog appears, enter a new duration in the Timecode field, and click Change. A window for changing the duration of a clip in the Timeline Ability to Delete Multiple Timeline Gaps at Once You can now ripple-delete video and audio gaps in the Timeline all at once using the Edit > Delete Gaps command. This removes gaps among consecutive clips in the Timeline on all Auto Select enabled tracks. Each segment of the Timeline with a gap is rippled, in order to move clips that are to the right of each gap left to close that gap. All gaps are defined for purposes of this command as empty spaces between clips that span all tracks in the Timeline. In the following example, various audio/video, audio-only, and video‑only clips have gaps between them. Using Remove Gaps causes the Timeline to be rippled such that these clips abut one another as a continuous sequence, without any of them overlapping any others. Before removing gaps Chapter – 2 Edit Page Improvements 1-26 After removing gaps This is an extremely powerful and wide-ranging command. However, it’s made safer by following strict rules in order to maintain overall A/V sync in timelines: Gaps will not be removed past the point where video and/or audio clips will overlap one another Gaps will not be removed if they’re under superimposed video clips that bridge the gap Gaps will not be removed if one or more continuous audio clips bridge the gap If a linked set of Video and Audio items has a gap that includes an L or J split edit, it will be closed to the point that the audio or video, whichever extends the farthest, abuts the nearest clip to it Disabling a track’s Auto Select Control omits that track from consideration when following the above rules. This lets gaps on other tracks be closed so clips overlap those on the Auto Select-disabled track WARNING: Performing Remove Gaps with Auto Select disabled on one or more tracks could result in massive loss of video/audio sync if you’re not careful. To avoid this, Shift-click one video Auto Select Control (or press Option-F9) and one audio Auto Select control (or press Command-Option-F9) to toggle all video and all audio Auto Select Controls until they’re all turned on at once. Improved Separation Between the Video and Audio Tracks The separator between the video and audio tracks in the Timeline has been made thicker and easier to drag. Improved Ripple Cut and Ripple Delete Behavior Performing a Ripple Cut or Ripple Delete upon multiple tracks using In and Out points ripples all Auto Select enabled tracks. Chapter – 2 Edit Page Improvements 1-27 Improved Automatic Audio Track Creation When dragging an audio clip to the undefined gray area of the Timeline below currently existing audio tracks in order to create a new track, the new track is set to a channel mapping that reflects the number of channels of the audio clip you’re dragging. This also means that if you’ve used Clip Attributes to map a clip’s audio to consist of multiple tracks where each track has a different channel mapping, for example, one 5.1 track, one stereo track, and six mono tracks, then editing that clip into the Timeline so that the audio portion creates new tracks will automatically create eight tracks, one that’s 5.1, one that’s stereo, and six that are mono. New Play Again Command The Play Again command (Option-L) lets you restart playback from where the playhead began without stopping, for instances where you quickly want to replay the beginning of what you’re listening to. Option to “Stop and Go To Last Position” A new option, Playback > Stop and go to last position, lets you choose whether or not to have the playhead return to where playback began whenever you stop. This option is most useful when editing audio, although it’s available any time. This option is also available when you right-click on the Stop button in the transport controls of any viewer. A contextual menu appears where you can turn “Stop and go to last position” on or off as the default behavior. Single Viewer Mode is Available in Dual Screen Layout The Single Viewer mode is now available when using the dual screen Edit page layout. Copy and Paste Timecode in Viewer Timecode Fields You can right-click on most Viewer timecode fields in the Media, Edit, and Color pages to choose Copy and Paste commands from a contextual menu for copying and pasting timecode values. The timecode value you’re pasting must be valid timecode, for example you can’t paste 0 hour timecode onto a 1 hour timeline. Right-clicking on a timecode field to use the Copy Timecode command Chapter – 2 Edit Page Improvements 1-28 Improved Timecode Entry When typing a combination of numbers and periods to enter a timecode value in the Edit page, whether to move the playhead or trim selected clips, the numbers you type are not converted into actual timecode until you press the Return or Enter key, making it easier to see what you’re entering. (Top) Entering timecode, (Bottom) The result Replace Edit Using Clips Already in the Timeline To facilitate workflows where multiple clips are stacked in the Timeline to manually track different takes or versions of stock footage, VFX clips, or other versionable media, there’s a method of drag-and-drop replace editing that copies the grade of the clip being replaced to the clip you’re replacing it with at the same time, so that newer versions of effects can inherit the same grade as the previous version of the effect being replaced. This only works for clips that have already been edited into the Timeline and that are superimposed (over or under) other clips in the Timeline, such as in the following screenshot. Be aware that this technique can also be used for multiple selected clips on the Timeline, to do several replace edits all at once. (Left) Before replace editing a clip in the Timeline, (Right) After Command-dragging a clip over one under it in the Timeline to replace edit the one below with the one above Chapter – 2 Edit Page Improvements 1-29 To replace edit one clip that’s stacked on the Timeline into another: 1 Select one or more clips that are already on the Timeline. Typically these will be clips that are superimposed over other clips. 2 Hold the command key down while dragging one superimposed clip on top of another to overwrite a clip and copy its grade to the clip you’re overwriting it with. NOTE: This won’t work with clips you’re editing into the Timeline from the Media Pool or Source Viewer. Option to Ripple the Timeline in Paste Attributes When using Paste Attributes to copy speed effects from one clip to another, a checkbox lets you choose whether or not the pasted speed effect will ripple the Timeline. Transition Categories in the Effects Library Transitions have been organized into logical categories to make it easier to find what you need. Categories include Dissolve, Iris, Motion, Shape, and Wipes. Linked Move Across Tracks The Timeline > Linked Move Across Tracks setting lets you change how linked video and audio items move in the Timeline when you drag them up and down to reorganize clips from track to track. Depending on the task at hand, one or the other behaviors might be more convenient, but no matter how you have this mode set, video/audio sync is always maintained when you move clips left and right. When Linked Move Across Tracks is enabled: (On by default) Dragging one of a linked pair of video and audio items up or down in the Timeline moves the linked item up or down as well. So, moving a video clip from track V1 to V2 results in its linked audio clip moving from track A1 to A2 as well. (Left) Before dragging video up one track, (Right) After Chapter – 2 Edit Page Improvements 1-30 When Linked Move Across Tracks is disabled: Dragging one of a linked pair of video and audio items up or down to another track in the Timeline only moves that one item, other linked item(s) remain in the same track. So, moving a video clip from track V1 to V2 leaves the audio clip in track A1, where it was originally. This makes it easy to reorganize video clips into different tracks while leaving your audio clips organized the way they were, or vice versa. Keep in mind that in this mode, while you can move one item of a linked pair up and down freely, moving that item left or right results in all linked items moving by the same amount, so sync is maintained. (Left) Before dragging video up one track, (Right) After Track Destination Keyboard Shortcuts Also Disable Tracks Pressing any of the Track Destination Selection keyboard shortcuts (Option-1–8 for video, Option-Control-1–8 for audio) repeatedly toggles the destination control on that track on and off. Ability to Mark In and Out Points in Cinema Mode When you’re in the full-screen playback of Cinema Mode on the Edit page, you can use the I and O keys to mark In and Out points on the Timeline. If you move the pointer, you can see marked timeline In and Out points on the Jog Bar of the onscreen controls, before they fade away again. Updated Timeline View Options Menu in the Toolbar The Timeline View Options menu has been updated with new options at the top to enable tabbed and stacked Timelines and to show and hide the Subtitle and Captions tracks area of the Timeline, while the third option is the previously available button to show or hide audio waveforms. The other previously available options are located below. Chapter – 2 Edit Page Improvements 1-31 The new Timeline View Options menu Edit Page Effects Enhancements A variety of Edit page effects features have been added and improved in DaVinci Resolve 15. The Text+ Title Generator in the Edit Page A new kind of title generator, named TextPlus, is available in the Titles category of the Effects Library’s Toolbox. This is the exceptionally fully-featured 2D text generator from Fusion, available for editing and customizing right in the Edit page. The new Text+ title generator, along with new Fusion titles below You can use the TextPlus generator the same way you use any generator in the Edit page. Simply edit it into a video track of the Timeline, select it, and open the Inspector to edit and keyframe its numerous properties to create whatever kind of title you need. Chapter – 2 Edit Page Improvements 1-32 In addition to having many more styling origins, the origin of the TextPlus generator in a compositing tool means that it offers many more panels worth of keyframable parameters, along with advanced animation controls built-in. These include keyframable Write On/Write Off controls, layout and animation using shapes (options include point, frame, circle, and path), character, word, and line transforms and animation, advanced shading, and full interlacing support. Four panels of the Text+ title generator, including Text, Layout, Transform, and Shading Chapter – 2 Edit Page Improvements 1-33 Better yet, with the playhead parked on your new TextPlus “Fusion Title,” you can open the Fusion page and access its parameters there too, if you want to start building upon this single generator to create a multi-layered motion graphics extravaganza. Opening the Text+ node in the Fusion page reveals it as an actual Fusion page operation Fusion Titles and Fusion Templates The abundance of other Fusion Titles in the Effects Library are custom-built text compositions with built-in animation that expose custom controls in the Inspector. A Fusion Title creating an animated lower third, with controls open in the Inspector In actuality, these text generators are Fusion templates, which are Fusion compositions that have been turned into macros and come installed with DaVinci Resolve to be used from within the Edit page like any other generator. It’s possible to make all kinds of Fusion title compositions in the Fusion page, and save them for use in the Edit page by creating a macro and placing it within the /Library/Application Support/ Blackmagic Design/DaVinci Resolve/Fusion/Templates/Edit/Titles directory, but this is a topic for another day. There’s one other benefit to TextPlus generators, and that is they can be graded like any other clip, without needing to create a compound clip first. Chapter – 2 Edit Page Improvements 1-34 Support for Caching of Titles and Generators You can now enable Clip caching for titles and generators, in case you’re using processor intensive effects. Position Curves in the Timeline Curve Editor You can now expose Position X and Position Y curves in the Edit page clip Curve Editor. This makes it possible to edit X and Y position keyframes independently, as well as to adjust the arc of geometric curves of the motion path in the Timeline Viewer by selecting one or more control points and making them into Bezier control points. Position X and Y curves in the Edit page Curve Editor Keyframable OpenFX and ResolveFX The parameters of OpenFX and ResolveFX have keyframe controls to the right of each parameter’s numeric field in the Inspector of the Edit and Color pages, so you can animate effects that you add to clips and grades. ResolveFX can now be animated in the Edit page using keyframe controls in the Inspector Chapter – 2 Edit Page Improvements 1-35 NOTE: There are no keyframe tracks in the Keyframe Editors of the Edit page or Color page at this time, so keyframes added to OpenFX or ResolveFX can only be edited in the Inspector. Compound Clips Support Alpha Transparency Creating compound clips preserves transparency of clips with alpha channels. For example, if you edit a series of title generators into the Timeline and turn them into a compound clip, the resulting compound clip will have the same transparency as the original generators. Improved Smooth Cut The Smooth Cut transition lets you seamlessly remove certain kinds of small jump cuts in interview clips. This transition has been updated with a Mode pop-up menu giving two options: Faster and Better. The Faster option is the original Smooth Cut method, which morphs between stills of the outgoing and incoming frames. The Better option is the new default, with improved quality and the capability of preserving the motion of subjects over the life of the transition. In most practical circumstances, the Better mode will give you a superior result, but certain cuts or effects may be better addressed with the Faster option. Improved Optical Flow for Speed Effects There are additional “Enhanced” Optical Flow settings available in the “Motion estimation mode” pop-up in the Master Settings panel of the Project Settings. These new settings provide better quality for slow motion and frame rate retiming effects, at the expense of being more processor intensive to play and render. The “Standard Faster” and “Standard Better” settings are the same options that have been available in previous versions of DaVinci Resolve. They’re more processor-efficient, and yield good quality that are suitable for most situations. However, “Enhanced Faster” and “Enhanced Better” should yield superior results in nearly every case where the standard options exhibit artifacts, at the expense of being more computationally intensive, and thus slower on most systems. Chapter – 2 Edit Page Improvements 1-36 Chapter 3 Subtitles and Closed Captioning DaVinci Resolve 15 adds new features to support subtitles and closed captioning in sophisticated ways. With dedicated subtitle/closed caption tracks that can be shown or hidden, subtitle file import and export, sophisticated subtitle editing and styling at the track and clip level, and comprehensive export options, adding subtitles and closed captions to finish your project is a clear and straightforward workflow. Chapter – 3 Subtitles and Closed Captioning 1-37 Contents Subtitles and Closed Captioning Support 1-39 Viewing Subtitle/Caption Tracks 1-39 Importing Subtitles and Captions 1-40 Adding Subtitles and Captions Manually 1-42 Editing Subtitles and Captions 1-44 Styling Subtitles and Captions 1-44 Linking Subtitles to Clips 1-45 Exporting Subtitles and Closed Captions 1-46 Naming Subtitle Tracks 1-47 Chapter – 3 Subtitles and Closed Captioning 1-38 Subtitles and Closed Captioning Support Subtitles are supported in DaVinci Resolve using specially typed subtitle tracks containing specifically designed subtitle generators to add and edit subtitles for a program. Typically each subtitle track corresponds to a single language or use, and you can change the name of a subtitle track to reflect its contents. Subtitle tracks can be locked, have auto select controls, and can be enabled or disabled like any other track. Additionally, a special subtitle-only destination control lets you choose which subtitle track to edit subtitle clips into. Furthermore, subtitle generator clips can be resized, moved, edited, and overwritten like most other clips. A subtitle track in the timeline Viewing Subtitle/Caption Tracks One important difference between subtitle tracks and other kinds of tracks is that only one subtitle track can be visible at any given time. That means if you have multiple subtitle tracks, each for a different language, clicking the Enable control for one subtitle track disables all others. Viewing one subtitle track at a time Chapter – 3 Subtitles and Closed Captioning 1-39 Importing Subtitles and Captions Oftentimes, adding subtitles or closed captions to a DaVinci Resolve timeline will involve importing a subtitle file that’s been prepared elsewhere. Currently, DaVinci Resolve supports subtitle files in the .srt SubRip format. To import an .SRT-formatted subtitle or closed captioning file: 1 Open the Media Pool. 2 Right-click on any bin in the Bin list, or anywhere in the background of the Media Pool browser, and choose Import Subtitle. 3 In the resulting file dialog, find and select the subtitle file you want to import, and click Open. 4 The subtitle file appears as a subtitle clip in the Media Pool, ready for editing into a subtitle track. An “ST” badge indicates that it’s a subtitle clip. An imported .srt subtitle file 5 To add a subtitle clip to a timeline, do one of the following: Drag a subtitle file you’ve imported into the unused gray area at the top of your video tracks, and a subtitle track will automatically be created for adding those subtitles into Drag a subtitle file you’ve imported into a pre-existing subtitle track As you drag the subtitle clip, it’ll immediately be decomposed so that each title is added to the Timeline as an individual subtitle clip, with its timing offset relative to the position of the first frame of the first subtitle in that file. (Top) The original timeline, (Bottom) The timeline after dragging a subtitle file has created a new subtitle track Chapter – 3 Subtitles and Closed Captioning 1-40 6 Position the imported subtitles so that they align with the first frame of your program that they’re supposed to, and drop the titles into the track. If you inadvertently misplace the subtitles, don’t worry, you can always select them all and slide them earlier or later, just like any other clips. 7 If you’ve added a new subtitle track, you can rename it to identify what language and country that track corresponds to. Please note that subtitle track names are used when exporting or encoding subtitles, so please make sure your tracks are named appropriately prior to export/delivery. 8 If you want to restyle all of the subtitles you’ve just added, for example to make them smaller or change the font, then click on the header of the subtitle track you’ll be working on, open the Track Style panel of the Inspector, and select the formatting you want that track to use. To see a list of every subtitle clip you’ve added, you can select the header of the subtitle track you’ve just added and open the Captions panel in the Inspector. A list at the bottom of the Captions panel gives you a convenient way of navigating the subtitles in a given track (using the Prev and Next buttons) and making selections. If you set the Inspector to be full height, you’ll have even more room for browsing the subtitle list. The Captions list shows you every caption or subtitle on a track, for selecting, editing, deleting, or navigating Chapter – 3 Subtitles and Closed Captioning 1-41 Adding Subtitles and Captions Manually Other times, you may need to create subtitles on your own. Before doing so, you’ll need to add one or more subtitle tracks. Once those tracks are created, you can add subtitle generators to them in a variety of ways. You can add as many subtitle tracks as you need, one for each language you require. To add new subtitle tracks: Right-click in any track header of the currently open timeline, and choose Add Subtitle Track. An empty subtitle track will appear at the top of the Timeline, named “Subtitle 1.” Once you’ve added a new subtitle track, you can rename it to identify what language and country that track corresponds to. Please note that subtitle track names are used when exporting or encoding subtitles, so please make sure your tracks are named appropriately prior to export/delivery. Showing and hiding subtitles tracks: Open the Timeline View options, and click on the Subtitle button to toggle the visibility of subtitles tracks on and off. The show/hide subtitle tracks button in the Timeline View Options To add individual subtitles to a subtitle track: 1 If you want to adjust the default style of a particular subtitle track before you start adding subtitles, then click on the header of the subtitle track you’ll be working on, open the Track Style panel of the Inspector, and select the formatting you want that track to use. 2 If you have multiple subtitle tracks, click the destination control of the subtitle track you want to add titles to. They’re labeled ST1, ST2, ST3, etc. 3 Move the playhead to the frame where you want the new subtitle to begin. Positioning the playhead where you want a new subtitle to begin Chapter – 3 Subtitles and Closed Captioning 1-42 4 To add a new subtitle clip, do one of the following: Open the Inspector and click Create Caption in the Captions panel of the Inspector. If there’s already one or more captions in that subtitle track, click the Add New button above the caption list, instead. Right-click anywhere on the subtitle track and choose Add Subtitle to add a subtitle clip starting at the position of the playhead Open the Effects Library, click the Titles category, and drag a Subtitle generator to the Subtitle track you want it to appear on. Manually adding a subtitle 5 If necessary, you can now edit the clip to better fit the dialog that’s being spoken or the sound that’s being described, by dragging the clip to the left or right, or dragging the beginning or end of the clip to resize it. 6 While the new subtitle clip you’ve created is selected, use the Captions panel in the Inspector to type the text for that particular subtitle. The text appears on the subtitle clip as you type it. Editing the text of the subtitle we just created Every time you add a subtitle, an entry is added to the subtitle list at the bottom of the Captions panel in the Inspector. This list gives you another convenient way of navigating the subtitles in a given track (using the Prev and Next buttons) and making selections. Chapter – 3 Subtitles and Closed Captioning 1-43 Editing Subtitles and Captions Subtitle clips can be selected singly or together, and slipped, slid, resized, rolled, and rippled just like any other clip in the Timeline, using the mouse or using keyboard commands, with either the Selection, Trim, or Razor tools. You can select subtitle clips in their entirety, or just their edit points, in preparation for nudging or dynamic trimming. In short, subtitle clips can be edited, in most ways, just like any other clips. Styling Subtitles and Captions When it comes to styling subtitle text, there are a wealth of styling controls in the Track Style panel of the Inspector. To modify the styling of all titles on a particular subtitle track: 1 Click on the header of the subtitle track you’ll be working on, or select a clip on a particular subtitle track either in the subtitle track or in the subtitle list of the Captions panel in the Inspector. 2 Open the Inspector, and then open the Track Style panel that appears within. 3 Edit whatever parameters you need to set the default style of all subtitles and closed captions that appear on that track. The Track Style panel has many more options than the Captions panel, including a group of Style and Position controls over Font and Font Face, Color, Size, Line Spacing, and Kerning, Alignment, Position X and Y, Zoom X and Y, Opacity, and Text Anchoring. The Track Style panel of the Inspector sets styling for every subtitle on that track Keep in mind that there are additional groups of controls that let you add a Drop Shadow, Stroke, and/or Background to all text on that track, which can be found at the bottom of the Track Style panel of the Inspector. Chapter – 3 Subtitles and Closed Captioning 1-44 Linking Subtitles to Clips If you like, you can link one or more subtitles to their accompanying clip, so that if you re-edit a subtitled scene, each clip’s subtitles move along with the clips. This arrangement doesn’t always work the way you’d expect when trimming, but it works great when you’re rearranging clips. To link a subtitle to another clip: 1 Select a clip and its subtitles all at once. Selecting a video clip and its accompanying subtitle to link them 2 Choose Clip > Linked Clips (Option-Command-L). A Link icon appears to show that the subtitle clips are linked to the video/audio clip. The now linked clip and subtitle have link badges to show their state Chapter – 3 Subtitles and Closed Captioning 1-45 Exporting Subtitles and Closed Captions Once you’ve set up one or more subtitle tracks in a program, the Deliver page exposes a group of Subtitle Settings at the bottom of the Video panel of the Render Settings that control if and how subtitles or closed captions are output along with that timeline. Available options for exporting subtitles can be found at the bottom of the Video panel of the Render Settings This panel has the following controls: Export Subtitle checkbox: Lets you enable or disable subtitle/closed caption output. Format pop-up: Provides four options for outputting subtitles/closed captions. –– As a separate file: Outputs each subtitle track you select as a separate file using the format specified by the Export As pop-up. A set of checkboxes lets you choose which subtitle tracks you want to output. –– Burn into video: Renders all video with the currently selected subtitle track burned into the video. –– As embedded captions: Outputs the currently selected subtitle track as an embedded metadata layer within supported media formats. There is currently support for CEA-608 closed captions within MXF OP1A and QuickTime files. You can choose the subtitle format from the Codec pop-up that appears. Export As: (only available when Format is set to “As a separate file”) Lets you choose the subtitle/closed captioning format to output to. Options include SRT and WebVTT. Include the following subtitle tracks in the export: (only available when Format is set to “As a separate file”) A series of checkboxes lets you turn on which subtitle tracks to output. Codec: (only available when Format is set to “As embedded captions”) Lets you choose how to format embedded closed captions; choices include Text and CEA-608. NOTE: Neither analog (Line 21) nor digital (CEA-708) closed caption output via Decklink or UltraStudio is supported at this time. Chapter – 3 Subtitles and Closed Captioning 1-46 Naming Subtitle Tracks If necessary, you can double-click the name of any subtitle track to rename it to something more descriptive of what that subtitle track will contain, such as the language, and whether a particular track is for subtitles or closed captions. Depending on your workflow and delivery specifications, there are existing conventions for identifying languages, such as ISO-639-1 (governing 2-letter codes) or ISO-639-2/B (governing 3-letter codes). These codes can be found at the International Organization for Standardization website, at http://www.loc.gov/standards/iso639-2/php/code_list.php. Some naming conventions require both language code and country code. For example, Facebook requires SubRip (.srt) files with the naming format “VideoFilename.[language code]_ [country code].srt” for proper embedding. If you want to use these codes for subtitle track identification and output, here’s a representative list of standardized language and country codes from around the world, in alphabetical order: Country ISO 639-1 Language Code ISO 639-2 Language Code ISO 3166-1 Country Code Amharic am amh ET (Ethiopia) Arabic ar ara EG (Egypt) AE (United Arab Emirates) LB (Lebanon) Bengali bn Chinese zh Danish da Dutch nl English en ben IN (India) CN (China) chi (B) zho (T) dan HK (Hong Kong) TW (Taiwan) DK (Denmark) dut (B) NL (Netherlands) nld (T) GB (UK) eng IN (India) US (US) Finnish fi fin FI (Finland) French fr fre (B) CA (Canada) fra (T) FR (France) German de Greek Modern el Hausa ha hau Hebrew he heb IL (Israel) Hindi hi hin IN (India) ger (B) deu (T) gre (B) DE (Germany) GR (Greece) ell (T) NG (Nigeria) TD (Chad) Chapter – 3 Subtitles and Closed Captioning 1-47 Country ISO 639-1 Language Code ISO 639-2 Language Code ISO 3166-1 Country Code Indonesian id ind ID (Indonesia) Italian it ita IT (Italy) Japanese ja jpn JP (Japan) Malay ms Maori mi mao (B) mri (T) NZ (New Zealand) Norwegian no nor NO (Norway) Polish pl pol PL (Poland) Portuguese pt por Punjabi pa pan IN (India) Russian ru rus RU (Russia) Spanish Castilian es spa may (B) msa (T) MY (Malaysia) BR (Brazil) PT (Portugal) CO (Columbia) ES (Spain) MX (Mexico) Swahili sw swa KE (Kenya) Swedish sv swe SE (Sweden) Tagalog tl tgl PH (Philippines) Thai tl tgl TH (Thailand) Turkish tr tur TR (Turkey) Urdu ur urd PK (Pakistan) Vietnamese vi vie VN (Vietnam) Chapter – 3 Subtitles and Closed Captioning 1-48 Chapter 4 Color Page Improvements DaVinci Resolve 15 is a great release for colorists. Timesaving new workflow features such as multiple playheads, the LUT browser, Timeline Grades browsers, and Shared Nodes make grade management faster than it’s ever been. New Matte Finesse controls, improved noise reduction, and camera raw controls for additional formats give you even more control and quality in everyday operations. Numerous Node Editor enhancements make it easier to see what’s happening in your grade. And finally, additional HDR tools supporting Dolby Vision and HDR10+ keep you at the cutting edge of grading and finishing. Chapter – 4 Color Page Improvements 1-49 Contents Clip, LUT, and Grade Browsing Features 1-51 Media Pool in Color Page 1-51 Dedicated LUT Browser 1-51 New Split Screen Modes to Preview Selected LUTs, Albums 1-53 Live Previews of Gallery Stills and LUTs 1-53 Live Previews of Composite Modes in the Layer Mixer 1-54 Live Previews of LUTs in the Node Editor 1-54 Favorite LUTs Submenu in Node Editor 1-54 Browse All Timeline Grades From the Current Project in the Gallery 1-54 Browse and Import Timeline Grades From Other Projects 1-55 RED SDK-Based RED IPP2 Setting in RCM Gamut Mapping 1-55 New Color Page Features 1-55 Multiple Timeline Playheads 1-55 Batch Version Management 1-56 Draggable Qualifier Controls 1-56 Additional Matte Finesse Controls 1-57 Node-Specific Color Space and Gamma Settings 1-57 Timeline Wipe Ganging to Current Clip 1-58 Camera RAW Palette for Canon RAW, Panasonic RAW Clips 1-58 Stereoscopic Convergence Support for Power Windows 1-58 Marker Overlays Visible in the Color Page Viewer 1-59 Timeline Marker List Available in Color Page Viewer Option Menu 1-59 Improved Noise Reduction 1-60 Improved Face Refinement 1-60 Node Editor Enhancements 1-60 Shared nodes for Group Management 1-60 Single-Click Node Selection in the Color Page 1-62 Ability to Disable/Enable Multiple Selected Nodes All at Once 1-62 Ability to Drag Connections From One Input to Another 1-63 Single-Click Connection Deletion 1-63 Edit Node Label Text by Double-Clicking 1-63 Dynamic Keyframe Indicator in the Node Graph 1-63 Improved Node Graph Interface 1-63 Thumbnail-Optional Nodes 1-64 Ability to Open Compound Nodes in “Display Node Graph” 1-64 HDR Enhancements 1-65 GPU-Accelerated Dolby Vision™ CMU Built-In (Studio Only) 1-65 Optional HDR10+™ Palette (Studio Only) 1-66 Chapter – 4 Color Page Improvements 1-50 Clip, LUT, and Grade Browsing Features A family of major new features let you work more efficiently with clips you want to use as external mattes, LUTs, and grades in the Gallery and in timelines of the current project or others. Media Pool in Color Page The Media Pool is available in the Color page, making it easy to drag and drop clips you want to use as External Mattes right into the Node Editor, for easy and fast connection to create various Color page effects. When opened, the Media Pool replaces the Gallery, fitting into the same area. In most respects, the Media Pool in the Color page works the same as the Media Pool on nearly every other page of DaVinci Resolve. The Media Pool now appears in the Color page When you drag a clip from the Color page Media Pool to the Node Editor, two things happen: That clip is turned into an External Matte in the current grade, which you can use as a Matte for secondary adjustments, or as a compositing layer (in conjunction with the Layer mixer) for mixing textures or images with your grade. That clip is also automatically attached to the Media Pool clip that corresponds to the clip you’re grading as a clip matte, to help you keep track of which clips are using other clips as mattes. Dedicated LUT Browser The LUT Browser provides a centralized area for browsing and previewing all of the LUTs installed on your workstation. All LUTs appear in the sidebar, by category. Chapter – 4 Color Page Improvements 1-51 The LUT Browser By default, all LUTs appear with a test thumbnail that give a preview of that LUT’s effect, but you can also get a Live Preview of how the current clip looks with that LUT by hover scrubbing with the pointer over a particular LUT’s thumbnail (this is described in more detail below). To open the LUT Browser: Click the LUT Browser button in the UI Toolbar at the top of the Color page. Methods of working with the LUT Browser: To see the LUTs in any category: Click on a LUT category to select it in the sidebar, and its LUTs will appear in the browser area. To make a LUT a favorite: Hover the mouse over a LUT and click the star badge that appears at the upper right-hand corner, or right-click any LUT and choose Add to Favorites. That LUT will then appear when you select the Favorites category. To search or filter for specific LUTs: Open a bin that has the LUT you’re looking for, then click the magnifying glass icon to open the search field, and type text that will identify the LUTs you’re looking for. To see LUTs in Column or Thumbnail view: Click the Column or Thumbnail buttons at the top right of the LUT Browser to choose how to view LUTs in the browser area. To sort LUTs in Thumbnail view: Click the Thumbnail Sort pop-up menu and choose which criteria you want to sort LUTs by. The options are Filename, Type, Relative Path, File Path, Usage, Date Modified. There are also options for Ascending and Descending sort modes. To sort LUTs in Column view: Click the column header to sort by that column. Click a header repeatedly to toggle between Ascending and Descending modes. To update the thumbnail of a LUT with an image from a clip: Choose a clip and frame that you want to use as the new thumbnail for a particular LUT, then right-click that LUT and choose Update Thumbnail With Timeline Frame. To reset the thumbnail of a LUT to use the standard thumbnail: Right-click a LUT and choose Reset Thumbnail to go back to using the standard test image. To refresh a LUT category with new LUTs that may have been installed: Select a LUT category, then right-click anywhere within the browser area and choose Refresh to refresh the contents of that category from disk. Chapter – 4 Color Page Improvements 1-52 Methods of adding LUTs to a grade from the LUT Browser: To apply a LUT to a clip: Select a clip in the Thumbnail timeline, then right-click a LUT and choose Apply LUT to Clip from the contextual menu. That LUT is added to the source clip, not the grade. To Append a LUT to the end of the node tree: Right-click a LUT and choose Append to Node Graph. A new node will be appended to the end of the current node tree with that LUT applied to it. To apply a LUT to a specific node: Drag a LUT from the LUT Browser and drop it onto the node you want to apply a LUT to. If you drag a LUT onto a node that already has a LUT, the previous LUT will be overwritten by the new one. New Split Screen Modes to Preview Selected LUTs, Albums A new Split Screen mode lets you simultaneously display previews of different LUTs as a split screen in the Viewer. To use this, turn Split Screen on, set the mode pop-up menu to LUTs, and then use the LUTs browser to Command-click up to 16 LUTs to preview in a grid. The Selected LUTs split screen option lets you preview a bunch of LUT looks at once Another new Split Screen mode lets you display every still within an album in the Gallery as a split screen in the Viewer. To use this, turn Split Screen on, set the mode pop-up menu to Album, and then open the Gallery and select an album. Only up to 16 stills will be displayed. Live Previews of Gallery Stills and LUTs The Live Preview option, found in both the Gallery and LUTs browser option menu, lets you preview how the current clip would look with any Gallery Still’s grade, or with any LUT applied to it, simply by moving the pointer over the still or LUT you want to preview. By default, the preview shows how the current clip would look if the scrubbed still or LUT replaced the grade currently applied to that still. The Live Preview option for the Gallery and LUT browser lets you hover over a LUT or saved grade to preview it on the current clip in the Viewe Chapter – 4 Color Page Improvements 1-53 Live Previews of Composite Modes in the Layer Mixer Scrolling through the submenu of Composite modes in a Layer Mixer node’s contextual menu now gives you a live update in the Viewer of how each transfer node affects the image. Live Previews of LUTs in the Node Editor If you hold down the Option key while scrolling through the submenu of LUTs in a Corrector node’s contextual menu, you’ll get a live update in the Viewer of how each LUT affects the image. Favorite LUTs Submenu in Node Editor When you label a LUT as a favorite in the LUT Browser, those favorite LUTs appear in a submenu of the contextual menu that appears when you right-click on a node in the Node Editor. This makes it easy to create a shortlist of your go-to LUTs for various situations, for rapid application right in the Node Editor. A Favorite LUTs submenu in the Node Editor contextual menu gives you a short list Browse All Timeline Grades From the Current Project in the Gallery The Gallery has a Timelines Album, available at the bottom of the Album list, that lets you browse all the grades in the current timeline, or in other timelines of the current project (using a pop-up menu that appears at the top of the Gallery browser area), making it easy to copy grades from earlier or later in your timeline, or from other timelines that share the same media. This is particularly useful for reality shows or documentaries where the same clips can appear multiple times in different parts of a program. Being able to simply show all existing grades in the Gallery frees you from having to save a still for every grade you think you might eventually reuse. The Timelines grade browser in the Gallery automatically shows all grades in the current timeline Chapter – 4 Color Page Improvements 1-54 Browse and Import Timeline Grades From Other Projects The Gallery Window lets you see and import grades in the timelines of other projects, even if they weren’t saved as stills first. When you open the Gallery Window and use the hierarchical disclosure controls of the Stills panel to open up and select a specific Database > User > Project > Timeline, you’ll see at least three browsable albums to the right, the Stills galleries that were created, the Memories, and at the bottom an album called Timeline Media. The Timeline Media album lets you browse the currently used grades for every clip in that timeline, making it easy to copy the ones you need to the current project’s Stills album or Memories. This is particularly useful if you’re working on a series where and you find that you want to re-use different grades, looks, adjustments, or fixes from previous episodes in the current one. Previously, you’d have to remember to save every clip as a still to be able to browse the grades in this way. Now you can simply browse the clips in the Timeline directly. Browsing the timeline grades for another project in the database RED SDK-Based RED IPP2 Setting in RCM Gamut Mapping RED WideGamutRGB and Log3G10 are now options in Resolve Color Management workflows using gamut mapping, and in the Gamut Mapping ResolveFX plug-in, to better support RED IPP2 workflows. New Color Page Features Many, many new general Color page features have been added to improve a wide variety of workflows Multiple Timeline Playheads DaVinci Resolve supports creating up to four separate playheads in the Mini-Timeline, that you can use to jump back and forth among different parts of your timeline. Only one playhead can be selected at any given time, and the currently selected playhead corresponds to the current clip, highlighted in orange. Each playhead in the Mini-Timeline is labeled with a letter, A through D. Chapter – 4 Color Page Improvements 1-55 Multiple playheads in the Mini Timeline for multi-region navigation To add a new playhead to the Mini-Timeline: Choose a playhead from the Color > Active Playheads submenu. That playhead will be placed at the same position as the original playhead, but it is now the one that is selected, so dragging the new playhead to a new position of the Mini-Timeline will reveal the original playhead you were using. To select another playhead to view: Click on the top handle of any playhead to select it, making that the currently active playhead controlled by the transport controls. By default, no keyboard shortcuts are mapped to the four playheads that are available, but you can create a custom keyboard mapping that you can use to quickly switch among them. Using the DaVinci Advanced Control Panel, you can use the A, B, C, and D buttons on the Jog/Shuttle panel to switch to the playhead you want to control. To eliminate all additional playheads from the Mini-Timeline: Choose Color > Active Playheads > Reset Playheads. Batch Version Management You can select multiple clips in the Thumbnail timeline and change them all to use a different Local or Remote version at once by right-clicking one of the selected thumbnails and choosing Load from the submenu of the Local or Remote version you want to switch to. Draggable Qualifier Controls The Qualifier controls now have draggable overlays, for more direct adjustment by mouse and tablet users. Drag the left and right edges of any qualifier control overlay to adjust the Low and High values (or the Width value of the Hue control). Drag the center of any qualifier control to change the center or to simultaneously change the Low and High parameters together. Optiondrag the left and right edges of any qualifier overlay to adjust softness. Draggable qualifier controls Chapter – 4 Color Page Improvements 1-56 Additional Matte Finesse Controls Denoise has been re-added to the Matte Finesse controls, providing a distinct way to postprocess extracted keys to selectively reduce the noise in a key, getting rid of stray areas of qualification and softly filling “holes” in a matte. Denoise in page 1 of the Matte Finesse controls An additional page of Matte Finesse controls expose controls for Shrink, Grow, Opening, and Closing functionality, with control over the Shape of the operation, the Radius, and Iterations. The previously available Black Clip and White Clip controls have also been moved into this second page. Shrink/Grow controls in page 2 of the Matte Finesse controls Node-Specific Color Space and Gamma Settings While the ability to change the color space a particular node works within has been available for several versions, the list of available color spaces has been greatly expanded (all the previous options such as Lab, HSL, and YUV are still there). Additionally, you have the option of choosing the gamma that node works with as well, with a similarly long list of options. Choosing a node-specific color space and gamma does not directly alter the image, as with the Color Space Transform ResolveFX plug-in. Instead, changing a node’s Color Space and Gamma alters what kind of channels the red, green, and blue controls affect, and how the adjustments you make within that node are applied. For example, this lets you apply OFX with Gamma set to Linear, which in some instances may be advantageous Chapter – 4 Color Page Improvements 1-57 . Expanded submenus for choosing a color space and gamma for image processing within a node Timeline Wipe Ganging to Current Clip The “Gang timeline wipe with current clip” option, available from the Viewer option menu, lets you maintain the offset between the current clip and a timeline clip you’re wiping against when you move the current clip selection to other clips. If you’re not sure what a Timeline wipe is, it’s when you use the Wipe Timeline Clip command in the Thumbnail timeline (it’s found in the contextual menu when you right-click a clip other than the current clip) to wipe the current clip against another clip in the timeline, without needing to save a still first. When you turn a Timeline wipe on, the timeline wiped clip is outlined in blue. With this new option enabled, the offset between the timeline wiped clip and the current clip is maintained when you move the clip selection. When this option is disabled, the timeline wiped clip stays where it is regardless of what clip you select. Camera RAW Palette for Canon RAW, Panasonic RAW Clips Canon RAW and Panasonic Varicam RAW media now expose dedicated controls in the Camera RAW panel of the Project Settings, and in the Camera RAW palette of the Color page when Canon or Panasonic raw media is present in the Timeline. Stereoscopic Convergence Support for Power Windows The Color group of the General Options panel of the Project Settings has a new checkbox called “Apply stereoscopic convergence to windows and effects” that correctly maintains the position of a window that’s been properly placed over each eye when convergence is adjusted. You must turn on a checkbox in the Project Settings to enable stereo convergence for windows When this option is enabled, the Window palette displays an additional Transform parameter, “Convergence,” that lets you create properly aligned convergence for a window placed onto a stereoscopic 3D clip. Chapter – 4 Color Page Improvements 1-58 The Convergence control in the Transform section of the Window palette After placing a window over a feature within the image while monitoring one eye, you can enable Stereo output in the stereo 3D palette and use the Pan and Convergence controls to make sure that window is properly stereo-aligned over the same feature in both eyes. At that point, adjusting the Convergence control in the Stereo 3D palette correctly maintains the position of the window within the grade of each eye. A convergence-adjusted window in stereo Marker Overlays Visible in the Color Page Viewer If you part the playhead on top of a marker in the timeline of the Color page, that marker’s information now appears in a Viewer overlay, just like in the Edit page, making it easier to read notes and see . Timeline Marker List Available in Color Page Viewer Option Menu The Option menu of the Color page Viewer has a submenu that lists all Timeline Markers in the currently open timeline. This makes it easy to run down client notes Timeline markers list available for quick access in the Viewer’s Option menu Chapter – 4 Color Page Improvements 1-59 Improved Noise Reduction A significantly improved new “Best” option has been added to the Mode pop-up of the Spatial NR controls, that does a much better job of preserving image sharpness and detail when raising the Spatial Threshold sliders to eliminate noise. The improvement is particularly apparent when the Spatial Threshold sliders are raised to high values of 40 or above (although this varies with the image). At lower values, the improvement may be more subtle when compared to the “Better” mode, which is less processor intensive than the computationally expensive “Best” setting. The new “Best” mode also allows you to decouple the Luma and Chroma Threshold sliders for individual adjustment, something you can’t do in “Better” mode. Improved Face Refinement Instances of face keyer chatter and flickering eyebag removal have been improved when using the Face Refinement plug-in. Node Editor Enhancements A variety of improvements to node editing have been added to DaVinci Resolve 15, starting with powerful new Shared nodes for making linked adjustments within clip grades, and continuing with multiple visual upgrades to nodes in the Node Editor. Shared nodes for Group Management This is probably one of the biggest improvements added for working colorists. You can now turn individual Corrector nodes into “Shared nodes,” which can then be copied to multiple clips to enable linked adjustments right from within the clip grade. This means that the clip grade can freely mix both clip-specific nodes and Shared nodes all within the same node tree. This makes Shared nodes fast to use as there’s no need to create groups or switch to a group node tree to reap the benefits of linked adjustments among multiple clips. A grade with an unshared (at left) and shared node (at right), a badge indicates the shared node, which is also locked What Are Shared Nodes Good For? Shared nodes are similar to group grades, except that they don’t require grouping and can be added to any normal grade. Changes made to a Shared node are automatically rippled to all other instances of that node in the grades of other clips. Furthermore, you can add as many Shared nodes to a grade as you like, and you can arrange them in any order to control the order of the operations they apply. And, of course, you can intersperse them with ordinary Corrector nodes Chapter – 4 Color Page Improvements 1-60 Shared nodes are extremely flexible. For example, you can use Shared nodes to: Add a Color Space Transform Resolve FX or a LUT to the beginning of every clip from a particular source Add a base correction to every talking head shot of a particular interviewee Add a shot matching adjustment to each clip from a particular angle of coverage within a scene Add a stylistic adjustment to every clip in a specific scene Use Shared nodes to make your base adjustments when grading with Remote Versions, so those adjustments remain linked when you copy your Remote Versions to Local Versions for fine tuning In fact, you can mix and match Shared nodes among differently overlapping sets of clips to accomplish any or all of the above at once. For example, you can add one Shared node to make an adjustment to every clip from a particular camera, add a second Shared node to each of those clips that are in a particular scene, and then add a third Shared node to whichever of those clips happen to be a closeup of the lead actress, before adding one or two regular Corrector nodes that aren’t shared to make clip-specific tweaks. IMPORTANT: At the time of this writing, there are two limitations when using Shared nodes. Grades using Shared nodes cannot use the Smart or User cache, and Shared nodes cannot be used in collaborative workflows. It is hoped that these limitations are temporary. Creating Shared Nodes Creating a shared node is easy, assuming you’ve created a node that has an adjustment you’d like to share among multiple clips. To create a Shared node: Right-click any Corrector node and choose “Save as Shared node.” Locking Shared Nodes Once you turn a node into a Shared node, that node is automatically locked, preventing you from accidentally making adjustments to it that would affect all other grades using that same Shared node. To toggle the locked status of a Shared node, do one of the following: Right-click any shared node and choose Lock Node from the contextual menu. Open the Keyframe Editor, and click the Lock icon in the track header of that node’s keyframe track. Copying Shared Nodes Because shared nodes are essentially Corrector nodes within of clip grades, they’re easy to work with. Once you’ve created one or more Shared nodes, there are a variety of ways you can copy them to the grades of other clips in your program to take advantage of the linked adjustments they let you make. Chapter – 4 Color Page Improvements 1-61 Ways of copying Shared nodes among multiple clips: Add a Shared node to another clip’s grade using the Node Editor contextual menu: Once you save a node as a Shared node, it becomes available from the bottom of the Add Node submenu of the Node Editor contextual menu, making it easy to add any Shared node to any clip. If you customize the label of the Shared node, that custom label appears in the contextual menu, making it easier to find what you’re looking for. Add Shared nodes to a basic grade you’ll be copying to other clips: If you create one or more Shared nodes when you initially build a grade, copying that grade to other clips naturally copies the Shared nodes as well. Save a Shared node as a Gallery still and apply it to other clips: If you save a grade with a Shared node in it to the Gallery, then every time you copy that Gallery still to another clip, you copy its Shared nodes. Create a Shared node and append it to a selection of additional clips: If you’ve already graded several clips in a scene, you can add a Shared node to the end of one of the clips grades and make sure it’s selected, then select all of the other clips in the scene and choose Append Node to Selected Clips. Use Shared nodes to preserve linked adjustments when copying Remote grades to Local grades: If you use Shared nodes to make your base adjustments when you grade using Remote Versions to automatically copy those grades to other clips that come from the same source media, those adjustments will remain linked when you copy your Remote Versions to Local Versions for fine tuning. Deleting Shared Nodes If you’ve created a Shared node that’s being used in multiple clips, and you decide you want to eliminate the linked relationship among these nodes so they all work independently, you can “delete” a specific shared node. This leaves the now unlinked nodes intact within each node tree in which they appear. Additionally, that Shared node is removed from the Add Node submenu. To Delete a Shared node: Right-click any Shared node, and choose a node to “un-share” from the Delete Shared Node submenu. Single-Click Node Selection in the Color Page You need only click once to select the current node in the Color page Node Editor, saving you years of wear and tear on your index finger. Ability to Disable/Enable Multiple Selected Nodes All at Once If you select more than one node in the node tree, using any of the available methods of turning nodes off and on (including Command-D) will toggle Enable/Disable Selected Nodes. This makes it easy to do before/after comparisons of any combination of nodes doing complicated adjustments by selecting them, while leaving un-selected nodes doing base adjustments that you want to leave enabled alone. Please note that the current node outlined in orange is always considered to be part of the selection. Chapter – 4 Color Page Improvements 1-62 Ability to Drag Connections From One Input to Another If you move the pointer over the second half of any connection line between two nodes, it highlights blue and you can click-and-drag it to reconnect it to another input. Single-Click Connection Deletion If you move the pointer over the second half of any connection line between two nodes so that it highlights blue, clicking once on the blue part of the connection deletes it. Edit Node Label Text by Double-Clicking Double-clicking the label of any node selects that label text for editing. This will only work if a node has a label already. Dynamic Keyframe Indicator in the Node Graph Nodes with keyframed parameters now display a keyframe badge in the Node Editor. Keyframed nodes now display a badge Improved Node Graph Interface The look and feel of nodes in the Node Graph has been updated for compatibility with the Fusion page Node Editor. Also, Mixer nodes have new icons to help identify them. Updated look and feel for Color page node trees Chapter – 4 Color Page Improvements 1-63 Thumbnail-Optional Nodes The Node Editor option menu provides a Show Thumbnails option that lets you disable or enable the thumbnails attached to each Corrector node. Disabling thumbnails in the Node Editor option menu makes nodes shorter Ability to Open Compound Nodes in “Display Node Graph” When you right-click a Gallery still or a thumbnail and choose Display Node Graph, you can now right-click compound nodes and choose “Show compound node,” or Command-double-click compound nodes to open them and see their individual nodes. Opening a compound node in a floating Node Graph window Chapter – 4 Color Page Improvements 1-64 HDR Enhancements DaVinci Resolve 15 adds support for the latest developments in HDR workflows. GPU-Accelerated Dolby Vision™ CMU Built-In (Studio Only) DaVinci Resolve 15 includes a GPU-accelerated software version of the Dolby Vision CMU (Content Mapping Unit) box for doing Dolby Vision grading and finishing workflows right in DaVinci Resolve. This is enabled and set up in the Color Management panel of the Project Settings with the Enable Dolby Vision checkbox. Turning Dolby Vision on enables the Dolby Vision palette in the Color page. Dolby Vision settings in the Color Management panel of the Project Settings Auto Analysis Available to All Studio Users Resolve Studio enables anyone to generate Dolby Vision analysis metadata automatically. This metadata can be used to deliver Dolby Vision content and to render other HDR and SDR deliverables from the HDR grade that you’ve made. Dolby Vision has two levels of metadata: i) Analysis metadata (Level 1), which is generated automatically by the project and image parameters; and ii) Artistic trim metadata (Level 2), which is set by the colorist to adjust the Dolby Vision mapped image to a target that is different from the mastering display (Rec. 709 as an example). Generating Dolby Vision with L1 analysis metadata is available without additional licensing from Dolby. Artistic trim metadata is created with the Dolby vision palette in DaVinci Resolve. The Dolby Vision Palette requires a separate license from Dolby. The Dolby Vision palette in the Color page Chapter – 4 Color Page Improvements 1-65 The commands governing Dolby Vision auto-analysis are in the Color > Dolby Vision™ submenu, and consist of the following commands: Analyze All Shots: Automatically analyzes each clip in the Timeline and stores the results individually. Analyze Selected Shot(s): Only analyzes selected shots in the Timeline. Analyze Selected And Blend: Analyzed multiple selected shots and averages the result, which is saved to each clip. Useful to save time when analyzing multiple clips that have identical content. Analyze Current Frame: A fast way to analyze clips where a single frame is representative of the entire shot. Analyze All Frames: A fast way to convert an entire HDR deliverable to a Dolby Vision deliverable without dealing with shot cuts. Manual Trimming Available Only to Licensees However, if you want to be able to make manual trims on top of this automatic analysis, email dolbyvisionmastering@dolby.com to obtain a license. Once you’ve obtained a license file from Dolby, you can import it by choosing File > Dolby Vision > Load License, and its successful installation will enable the Dolby Vision palette to be displayed in the Color page. Dolby Vision Metadata Export Additionally, DaVinci Resolve now has the ability to render MXF files with embedded Dolby Vision trim metadata. Optional HDR10+™ Palette (Studio Only) DaVinci Resolve 15 supports the new HDR10+ HDR format by Samsung. Please note that this support is a work in progress as this is a new standard. When enabled, an HDR10+ palette exposes trimming parameters that let you trim an automated downconversion of HDR to SDR, creating metadata to control how HDR-strength highlights look on a variety of supported televisions and displays. This is enabled and set up in the Color Management panel of the Project Settings with the Enable HDR10+ checkbox. Turning HDR10+ on enables the Dolby Vision palette in the Color page. Dolby Vision settings in the Color Management panel of the Project Settings HDR10+ Auto Analysis Commands HDR10+ has its own scheme for auto-analyzing HDR to SDR downconversion metadata, and the controls are available in the Color > HDR10+ submenu, consisting of the following commands: Analyze All Shots: Automatically analyzes each clip in the Timeline and stores the results individually. Analyze Selected Shot(s): Only analyzes selected shots in the Timeline. Analyze Selected And Blend: Analyzed multiple selected shots and averages the result, which is saved to each clip. Useful to save time when analyzing multiple clips that have identical content. Analyze Current Frame: A fast way to analyze clips where a single frame is representative of the entire shot. Chapter – 4 Color Page Improvements 1-66 HDR10+ Palette An HDR trim palette is available to all Resolve Studio users, that provides controls for manually trimming the auto-analyzed trim metadata. At the time of this writing, nine sliders control the bezier handles and control points of a custom curve that can be used to shape the luminance mapping curve, including Knee X and Y sliders. The Dolby Vision palette in the Color page HDR10+ Metadata Export The resulting metadata is saved per clip, in a JSON sidecar file. Chapter – 4 Color Page Improvements 1-67 Chapter 5 New ResolveFX DaVinci Resolve 15 adds seven new ResolveFX that will give colorists and finishing artists significant new image restoration tools, as well as sophisticated new lighting and stylistic effects with wide-ranging uses. Two exciting new plug-ins have been added to the “ResolveFX Light” category that simulate different types of optical glows and flares. Additionally, the ResolveFX category formerly called “Repair” has been renamed “ResolveFX Revival,” because of three powerful new restoration effects that have been added. On the other end of the spectrum, two new plug-ins that create stylistic degradation have been added to the “ResolveFX Texture” and “ResolveFX Transform” categories. Chapter – 5 New ResolveFX 1-68 Contents Aperture Diffraction (Studio Only) 1-70 Lens Reflections (Studio Only) 1-71 Patch Replacer (Studio Only) 1-73 Automatic Dirt Removal (Studio Only) 1-75 Dust Buster (Studio Only) 1-76 Deflicker (Studio Only) 1-77 Flicker Addition 1-79 Film Damage 1-79 Chapter – 5 New ResolveFX 1-69 Aperture Diffraction (Studio Only) Found in the “ResolveFX Light” category, Aperture Diffraction models the starburst effect usually seen when shooting bright lights with small apertures, the physical cause of which is light-diffraction on the aperture blades of a lens. This plug-in simulates this, with the result being automatically applied to scene highlights that you can isolate and refine, with customizable virtual apertures. Small regions of brightness exhibit a star pattern glow, as seen in the following image. (Left) Original image, (Right) Applying Aperture Diffraction Large regions of brightness exhibit a more even glow with shaping and texture that look like a natural optical effect. It can be used to create a different type of glow effect with a more realistic look in some situations than the Glow plug-in, though it’s more processor intensive. In other situations, this plug-in opens up many different stylistic possibilities for glowing effects. (Left) Original image, (Right) Applying Aperture Diffraction Output Select Output lets you preview the image with different stages of the Aperture Diffraction effect applied, viewing the Isolated Source (to help when adjusting the Isolation Controls), Preview Aperture (to help when adjusting the Aperture Controls), Preview Diffraction Pattern (showing you the resulting diffraction pattern based on the aperture control settings), Diffraction Patterns Alone (showing you the glow effect that will be applied to the image by itself), and the Final Composite. Isolation Controls The Isolation Controls control which highlights in the scene generate visible glow and patterns. The effect of these controls can be directly monitored by setting Select Output to Isolated Source. Color Mode is a pop-up menu that lets you either choose to keep the colors of the different highlight regions that generate glow, or treat them all as greyscale brightness only (color controls later can change the effect). Greyscale is faster to process, but Color can result in some brilliant effects. Chapter – 5 New ResolveFX 1-70 Brightness sets the threshold at which highlights are isolated. Gamma lets you shape the isolated highlights, while Smooth lets you blur details in the highlights that you don’t want to be pronounced. Color Filter lets you choose a particular color of highlight to isolate (an eyedropper lets you select a value from the Viewer). The Operation controls let you adjust the resulting Isolation matte (options include Shrink, Grow, Opening, Closing) with a slider to define how much. Aperture Controls The Aperture Controls let you define the shape and texture of the resulting glow this plug‑in creates. Iris Shape lets you choose a shape that determines how many arms the star pattern will have. Aperture size lets you alter the resulting diffraction pattern alternating between more of a star shape at higher values and a stippled wave pattern at lower values. Result Gamma lets you adjust how pronounced will be the glow that appears between the arms of the star patterns that appear. Result Scale lets you alternate between pronounced star patterns at high values and more diffuse glows at low values. Blade Curvature and Rotation let you alter the softness and orientation of the arms of each star. Chroma Shift lets you introduce some RGB “bleed” into the glow. Compositing Controls These controls let you adjust how to composite the glow effect against the original image. The Normalize Brightness checkbox scales the brightness of the glow to a naturalistic range for the image. Also, when Normalize Brightness is enabled, the Aperture Diffraction effect will keep to a consistent overall brightness as the scene changes. Brightness lets you adjust the intensity of the glow effect. Colorize lets you tint the glow effect using a Color control that appears when Colorize is raised above 0. Lens Reflections (Studio Only) Found in the “ResolveFX Light” category, Lens Reflections simulates intense highlights reflecting off of the various optical elements within a lens to create flaring and scattering effects based on the shape and motion of highlights you isolate in the scene. It’s an effective simulation that works best when there are light sources or specular reflections in the scene such as the sun, car headlights, light fixtures, fire and flame, or other lighting elements that are plausibly bright enough to cause such flaring. Also, this plug-in really shines when these light sources move, as each layer of simulated reflections moves according to that element’s position within the virtual lens being simulated, creating organic motion that you don’t have to keyframe. Without intense highlights, the results of this filter will be somewhat abstract. (Left) Original image, (Right) Applying Lens Reflections Chapter – 5 New ResolveFX 1-71 Output Select Output lets you preview the image with different stages of the Lens Reflections effect applied, viewing the Isolated Source (to help when adjusting the Isolation Controls), Reflections Alone (showing you the flaring effect that will be applied to the image by itself), and the Final Composite. A Quality pop-up lets you choose how to render the effect. Options are Full, Half (Faster), and Quarter (Fast). The tradeoff is between quality and speed. Isolation Controls The Isolation Controls control which highlights in the scene generate lens reflections. The effect of these controls can be directly monitored by setting Select Output to Isolated Source. It’s highly recommended to customize the Isolation controls for the image at hand when using this plug-in, as even more so than other plug-ins, the particular highlights used will have a huge impact on the resulting effect. Color Mode is a pop-up menu that lets you either choose to keep the colors of the different highlight regions that generate lens reflections, or treat them all as grayscale brightness only (color controls later can change the effect). Grayscale is faster to process, but Color can result in some brilliant effects. Brightness sets the threshold at which highlights are isolated. Gamma lets you shape the isolated highlights, while Smooth lets you blur details in the highlights that you don’t want to be pronounced. Color Filter lets you choose a particular color of highlight to isolate (an eyedropper lets you select a value from the Viewer). The Operation controls let you adjust the resulting Isolation matte (options include Shrink, Grow, Opening, Closing) with a slider to define how much. Global Controls The Global Controls let you quickly adjust the overall quality of the Lens Reflections effect. Global Brightness lets you raise and lower the level of all reflections. Global Blur lets you defocus all reflections. Anamorphism lets you deform the reflection elements to simulate an anamorphic lens’ stretching effect. Global Colorize lets you adjust the color intensity of the reflections, either intensifying the color of all reflections or desaturating it. Presets A Presets pop-up provides a number of different settings to get you started. Selecting a preset populates the Reflecting Elements parameters below, at which point you can customize the effect to work best with the image at hand. It’s highly recommended to customize these effects to suit the type of highlights in your image, in order to get the best results. Chapter – 5 New ResolveFX 1-72 Reflecting Elements There are four groups of Reflecting elements, each with identical controls. This lets you create interactions combining up to four sets of reflections. The controls found within each group are as follows. Brightness: Lets you adjust the intensity of that reflection. Position in Optical Path: Lets you shift the reflection according to an element’s position in the lens. Practically, this means that positive values will enlarge an inverted reflection based on the highlights, while reducing values towards 0 will shrink the reflection, and pushing this into negative values will invert the reflection and pull it into the opposite direction as it begins to enlarge again. A value of –1 positions the reflection right over the highlight that creates it. Defocus type: Lets you choose what kind of blur to use, choices include Box Blur, Triangular Blur, Lens Blur (the most processor intensive), and Gaussian Blur (the default). Defocus lets you choose how much to blur that element. Stretch: Lets you give the flare an anamorphic widescreen look, while Stretch Falloff lets you taper the edges. Lens Coating: A pop-up lets you choose common colors such as purple, green, and yellow that correspond to different anti-reflective lens coatings, as well as a selection of other vibrant colors. A color control and eyedropper let you manually choose a color or pick one from the image. A Colorize slider lets you vary how much to tint the reflection by the selected color, although setting Colorize to 0 lets the flare take its color from the source highlights of the image, which can sometimes give you the most interesting look. Patch Replacer (Studio Only) Found in the new “ResolveFX Revival” category, the Patch Replacer is a quick fix when you need to “paint out” an unwanted feature from the image. For those of you who’ve been using windows and Node Sizing to do small digital paint jobs, this plug-in offers more options and a streamlined workflow. On adding the plug-in, an onscreen control consisting of two oval patches appears, with an arrow connecting them indicating which patch is being copied into the other. The oval to the left is the “source” patch, used to sample part of the image, and the oval to the right is the “target” patch, used to cover up the unwanted feature using pixels from the source patch. To use the Patch Replacer, simply drag the target patch over the feature you want to obscure, resize it to fit using the corner controls (the source patch is automatically resized to match), and then drag the source patch to an area of the image that can convincingly be used to fill the target patch. (Left) Original image, (Right) Removing the thermostat with the Patch Replacer Chapter – 5 New ResolveFX 1-73 The source and target patches can be motion tracked using the FX tracker, so this tool is effective even if the subject or camera is moving. Main Controls The “Fill-in method” pop-up menu is arguably the most important, as it defines what method to use to fill the destination patch with whatever is in the source patch. The rest of the primary controls work differently depending on which fill-in method you choose. Clone: Simply copies the source patch into the target patch. When clone is selected, the Replacement Detail slider (which defaults to 1) lets you fade out the source patch, while Region Shape lets you choose a different kind of shape to use, and Blur Shape Edges lets you feather the edge of this operation, to more convincingly blend the source with the target area. Adaptive Blend: A much more sophisticated method of obscuring the target area using pixels from the source patch, and in many cases will yield better results more quickly than cloning. The source patch is copied into the target patch in such a way as to combine the source detail with the lighting found inside of the target area, creating in most instances a fast, seamless match. The Keep Original Detail checkbox, when turned on, merges detail from the source and target patches to create a composite, rather than a fill. The Blur Shape Edges slider works a bit differently with Adaptive Blend selected, but the idea is the same, feathering the effect from the outside in to obscure instances where there’s a noticeable border around the target area. Fast Mask: Eliminates the source patch, doing instead a quick neighboring pixel blend that works well with small patches, but can betray a grid pattern on larger patches. Region Shape and Blur Shape Edges are both adjustable. Patch Positions Source X and Y, Target X and Y, and Target Width and Height are provided as explicit controls both for numeric adjustment, should that be necessary, and also to allow for keyframing in case you need to change the position and/or size of the source and fill patches over time. Keep in mind that the source and target patches can be motion tracked using the FX tracker, although two checkboxes, Source Follows Track and Target Follows Track, let you disable FX tracker match moving when necessary. On-Screen Controls The Control Visibility pop-up menu lets you choose whether the source and target onscreen controls are visible as you work. Show (the default) leaves all onscreen controls visible all the time. Auto Hide hides all onscreen controls whenever you’re dragging one, letting you see the image as you adjust it without having these controls in the way. Hide makes all onscreen controls invisible, so you can see a clean version of the image with the effect, however you can still edit the effect if you remember where the controls are. Chapter – 5 New ResolveFX 1-74 Automatic Dirt Removal (Studio Only) Found in the new “ResolveFX Revival” category, the Automatic Dirt Removal plug-in uses optical flow technology to target and repair temporally unstable bits of dust, dirt, hair, tape hits, and other unwanted artifacts that last for one or two-frames and then disappear. All repairs are made while maintaining structurally consistent detail in the underlying frame, resulting in a high-quality restoration of the image. Fortunately, despite its sophistication, this is a relatively easy plug-in to use; just drop the plug-in on a shot, adjust the parameters for the best results, and watch it go. (Left) Original image, (Right) Using Automatic Dirt Removal NOTE: This plug-in is less successful with vertical scratches that remain in the same position for multiple frames, and is completely ineffective for dirt on the lens which remains for the entire shot. Main Controls Motion Estimation Type lets you choose from among None, Faster, Normal, and Better. This tunes the tradeoff between performance and quality. Neighbor Frames lets you choose how many frames to compare when detecting dirt. Choosing more frames of comparison take longer to process, but usually results in finding more dirt and artifacts. Repair Strength slider lets you choose how aggressively to repair dirt and artifacts that are found. Lower settings may let small bits through that may or may not be actual dirt, while higher settings eliminate everything that’s found. The Show Repair Mask checkbox lets you see the dirt and artifacts that are detected by themselves, so you can see the effectiveness of the results as you fine tune this filter. Fine Controls The Motion Threshold slider lets you choose the threshold at which pixels in motion are considered to be dirt and artifacts. At lower values more dirt may escape correction, but you’ll experience fewer motion artifacts. At higher values, more dirt will be eliminated, but you may experience more motion artifacts in footage with camera or subject motion. The Edge Ignore slider lets you exclude hard edges in the picture from being affected by dirt or artifacts that are removed. Higher values omit more edges from being affected. Chapter – 5 New ResolveFX 1-75 Dust Buster (Studio Only) Found in the new “ResolveFX Revival” category, this plug-in is also designed to eliminate dust, dirt, and other imperfections and artifacts from clips, but it does so only with user guidance, for clips where the Automatic Dirt Removal plug-in yields unsatisfactory results. This guidance consists of moving through the clip frame by frame and drawing boxes around imperfections you want to eliminate. Once you’ve drawn a box, the offending imperfection is auto-magically eliminated in the most seamless way possible. This works well for dirt and dust, but it also works for really big stains and blotches, as seen below. (Left) Drawing a box around dirt in the original image, (Right) Result in the Dust Buster plug-inn This plug-in works similarly to, but supersedes the legacy Dust Removal feature, which only worked on select image sequence formats, and wrote new media files on disk. The Dust Buster plug-in works on any format of movie clip, and works non-destructively, storing all image repairs within the plug-in without creating new media. Best of all, this plug-in is able to do its magic with only three controls. Mode: Selects how imperfections within the bounding box you draw are fixed. By default, Auto just takes care of things without you needing to think about this. However, if you’re not satisfied with the result, you can undo, and choose a different method from this pop-up method. Here are all the options that are available to you. –– Auto: The default method. Once you’ve drawn a bounding box, the two frames prior to and the two frames after the current clip will be analyzed and compared to the current image. The best of these 5 frames will be drawn upon to remove the imperfection in the current frame. Images two frames away are prioritized since that will avoid the appearance of frozen grain, but only if they’re suitable. –– Prev/Next Frame: If you draw a bounding box from left to right, the next frame will be drawn upon to remove the imperfection. If you draw a bounding box from right to left, the previous frame will be used. –– Prev–1/Next+1 Frame: If you draw a bounding box from left to right, the image two frames forward will be drawn upon to remove the imperfection. If you draw a bounding box from right to left, the image two frames back will be used. –– Spatial Fill: In cases where the other two modes yield unsatisfactory results, such as when the underlying image has fast or blurred motion, this mode uses surrounding information in the current frame to remove the imperfection. Show Patches: Off by default. Turning this checkbox on lets you see every bounding box you’ve drawn to eliminate imperfections. While the patches are shown, you can Shift-click to select individual patches, group select patches by Command-dragging a bounding box, and delete unwanted patches individually by Option-clicking them. Reset Frame: Resets all of the bounding boxes drawn on the current frame, so you can start over. Chapter – 5 New ResolveFX 1-76 Deflicker (Studio Only) Found in the new “ResolveFX Revival” category, this brand new plug-in replaces the previous “Timelapse Deflicker” filter, and solves a far broader variety of problems in a much more automatic way. The new Deflicker plug-in handles such diverse issues as flickering exposure in timelapse clips, flickering fluorescent lighting, flickering in archival film sources, and in certain subtle cases even the “rolling bars” found on video screens shot with cameras having mismatched shutter speeds. Two key aspects to this filter are that it only targets rapid, temporally unstable variations in lightness, and that it’s able to target only the areas of an image where flickering appears, leaving all other parts of the image untouched. As a result, this plug-in can often repair problems once considered “unfixable.” (Left) Original image with flicker, (Right) Result setting Deflicker to Fluro Light, (clip courtesy Redline Films) Main Settings By default, the top section of this plug-in exposes a single control, which in many cases may be all you need. Deflicker Setting pop-up menu: The top two options, Timelapse and Fluoro Light, are presets that effectively eliminate two different categories of flickering artifacts. If neither of these presets is quite as effective as you’d hoped, a third option, Advanced Controls, opens up the Isolate Flicker controls at the heart of this plug-in to let you tailor it further to your needs. Isolate Flicker Hidden by default, these controls only appear when you set “Deflicker Setting” to Advanced Controls, and let you choose how to detect motion in the scene so that flickering may be correctly addressed relative to the motion of subjects and items within the frame where it appears. Mo.Est. Type: Picks the method DaVinci Resolve uses to analyze the image to detect motion. Despite the names of the available options, which options will work best is highly scene dependent. The default, Faster, is less processor intensive, but less accurate, however this can be an advantage and actually do a better job with highdetail images that would confuse the Better option. Choosing Better is more accurate, but more processor intensive, and Better will try harder to match fine details which can sometimes cause problems. None lets you disable motion analysis altogether, which can work well (and will be considerably faster) in situations where there’s no motion in the scene at all. The default is Better. Frames Either Side: Specifies the number of frames to analyze to determine what’s in motion. Higher values are not always better, the best setting is, again, scene dependent. The default is 3. Chapter – 5 New ResolveFX 1-77 Motion Range: Three settings, Small, Medium, and Large, let you choose the speed of the motion in the frame that should be detected. Gang Luma Chroma: Lets you choose whether to gang the Luma and Chroma Threshold sliders or not. Luma Threshold: Determines the threshold above which changes in luma will not be considered flicker. The range is 0-100, 0 deflickers nothing, 100 applies deflickering to everything. The default is 100. Chroma Threshold: Determines the threshold above which changes in chroma will not be considered flicker. The range is 0-100, 0 deflickers nothing, 100 applies deflickering to everything. The default is 100. Motion Threshold: Defines the threshold above which motion will not be considered flicker. Speed Optimization Options Closed by default, opening this control group reveals two controls: Reduced-Detail Motion checkbox: On by default, reduces the amount of detail that’s analyzed to detect flicker. In many cases, this setting makes no visible difference, but increases processing speed. Disable this setting if your clip has fine detail which is being smoothed too aggressively. Limit Analysis Area checkbox: Turning this on reveals controls over a sample box that you can use to limit deflickering to a specific region of the image. This option is useful when (a) only one part of the image is flickering, so focusing on just that area speeds the operation considerably, or (b) part of the image is being smoothed too much by deflickering that’s fixing another part of the image very well. Restore Original Detail After Deflicker Closed by default, opening this control group reveals two controls: Detail to Restore slider: Lets you quickly isolate grain, fine detail, and sharp edges which should not be affected by the deflicker operation, preserving those fine details exactly. Show Detail Restored checkbox: Turning this checkbox on lets you see the edges that are detected and used by the Detail to Restore slider, to help you tune this operation. Output The output pop-up menu lets you choose what Deflicker outputs, with options to help you troubleshoot problem clips. Here are the available options: Deflickered Result: The final, repaired result. This is the default setting. Detected Flicker: This option shows you a mask that highlights the parts of the image that are being detected as having flickering, to help you evaluate whether the correct parts of the image are being targeted. This mask can be very subtle, however. Magnified Flicker: This options shows you an exaggerated version of the Detected Flicker mask, to make it easier to see what the Deflicker plug-in is doing. Chapter – 5 New ResolveFX 1-78 Flicker Addition On the other hand, why remove flicker when you can add it instead? Found in the “ResolveFX Transform” category, the Flicker Addition plug-in adds rapidly animated exposure changes to make the image appear to flicker, creating animated effects that would be difficult to keyframe manually. When applied to an image in different ways, this plug-in can be used to simulate torchlight, firelight, light fixtures with old ballasts or frayed wiring, or any temporally unstable light source. For example, you could key only the highlights of a night-time image, and use Flicker Addition to affect those isolated highlights. Two groups of controls let you control the quality of this flickering. Main Controls The Flicker Type pop-up menu lets you apply the flicker as a Lift, Gamma, Gain, or Vignette adjustment. The Range slider lets you set how widely the flickering will vary. Speed lets you adjust how quickly the flickering is animated. The Smoothness slider lets you adjust the temporal quality of the flickering, whether it changes abruptly from one value to another (at lower settings) or whether it makes more continuous transitions from one value to another (at higher settings). Three checkboxes let you choose which color channels are affected by this flickering. Flicker Quality These controls let you adjust the details of how the flickering animates. The Randomness Scale slider lets you introduce irregularity to the Horizontal, Vertical, and Rotational motion of the camera shake. The greater this value, the more irregularity will be introduced. The Randomness Speed slider lets you choose between smoothly erratic motion (at lower values) or more jagged motion (at higher values). The Pause Length slider lets you adjust the frequency of intermittent pauses that break up the random motion added by this filter. The Pause Interval slider lets you adjust the duration of intermittent pauses that break up the random motion added by this filter. The Pause Randomness lets you add a degree of randomness to the intervals that happen. The Random Seed slider lets you alter the value that sets what random values are being 1038 produced. Identical values result in identical randomness. Film Damage Found in the “ResolveFX Texture” category. After you’ve used the new ResolveFX Revival plug-ins to fix damage in archival footage, you can turn around and use the Film Damage plug-in to make brand new digital clips look worn, dirty, and scratched instead. When used in conjunction with the Film Grain and Flicker Addition plug-ins, you can convincingly recreate the feel of poorly kept vintage archival footage. (Left) Original image, (Right) Result with Film Damage Chapter – 5 New ResolveFX 1-79 Blur and Shift Controls The three parameters at the top let you alter the foundation of the image to begin creating the look of an older film. Film Blur lets you add just a bit of targeted defocusing to knock the digital sharpness out of the image. Temp Shift defaults to warming the image just a bit to simulate the warmer bulb of a film projector, although you can use it to cool or warm the image in varying amounts. Tint Shift defaults to yellowing the image to simulate damage to the film dyes, but you could move the slider in the other direction to add a bit of magenta, simulating a different kind of dye failure. Add Dirt These parameters let you simulate dirt particles (not dust) that have adhered to the film. These are larger specks, although there’s several ways you can customize these. The Dirt Color control lets you choose what color you want the dirt particles to be (black simulates dirt on a print, while white simulates dirt on a negative). The Changing Dirt checkbox lets you alternate between simulating temporally unstable dirt on the film (checkbox on), and dirt on the lens that doesn’t move (checkbox off). Dirt Density lets you choose more or less dirt particles appearing over time. Dirt Size lets you choose the average size of dirt particles to appear. Dirt Blur lets you defocus the dirt so it’s not so sharp. Dirt Seed changes the random distribution of dirt when you change this value, but for any given value, the results for any given set of control adjustments remain consistent. Add Scratch These parameters add a single scratch to the image, simulating something scratching the emulsion while the film played. Scratch Color lets you choose the color you want the scratch to be (scratches can be a variety of colors depending on the depth of the scratch, type of film, and method of printing). Scratch Position lets you adjust the scratch’s horizontal position on the image. Scratch Width and Scratch Strength let you adjust the scratch’s severity, while Scratch Blur lets you defocus it. The Moving Scratch checkbox lets you choose whether the scratch is jittering around or not. Moving amplitude determines how far it moves. Moving Speed determines how fast it moves. Moving Randomness determines how it meanders about, and Flickering Speed determines how much the scratch flickers lighter and darker in severity. Add Vignetting These parameters simulate lens vignetting darkening the edges of the image. Focal Factor adjusts how far the vignetting extends into the image. Geometry Factor affects how dark the vignetting is, and how pronounced the edges are. Tilt Amount affects how balanced the vignetting at the top of the image is versus the bottom of the image, while Tilt Angle affects how balanced the vignetting left of the image is versus the right, but only when Tilt Amount is set to something other than 0. Chapter – 5 New ResolveFX 1-80 Chapter 6 Fairlight Page Improvements The Fairlight Page is full of big new features and small refinements, making this an even more professional tool for audio postproduction, whether you’re doing dialog editing, sound design, audio cleanup, or mixing. With new tools for automated dialog replacement, sound effects searching and auditioning, VSTi support for samplers to do MIDI controller-driven foley, 3D audio panning, clip and track bouncing, and numerous new UI refinements, controls, and commands for audio playback and editing, there’s something for every kind of audio professional. Chapter – 6 Fairlight Page Improvements 1-81 Contents New Fairlight Page Features 1-83 ADR (Automated Dialog Replacement) 1-83 Fixed Playhead Mode 1-90 Video and Audio Scrollers 1-90 3D Audio Pan Window 1-91 Visible Video Tracks 1-92 User-Selectable Input Monitoring Options 1-93 Commands For Bouncing Audio 1-93 Sound Library Browser 1-94 VSTi Support For Recording Instrument Output 1-96 General Fairlight Page Enhancements 1-99 Normalize Audio Levels Command 1-99 Clip Pitch Control 1-99 Support for Mixed Audio Track Formats from Source Clips 1-100 Oscillator for Generating Tone, Noise, and Beeps 1-100 Compound Clips Breadcrumb Controls Below Fairlight Timeline 1-101 Level and Effects Appear in Inspector for Selected Bus 1-101 Audio Waveform Display While Recording 1-101 Audio Playback for Variable Speed Clips 1-101 Paste and Remove Attributes for Clips, Audio Tracks 1-102 Loop Jog Scrubbing 1-102 Improved Speaker Selection Includes Multiple I/O Devices 1-102 Fairlight Page Editing Enhancements 1-103 Media Pool Preview Player 1-103 Edit Page—Compatible Navigation and Selection Keyboard Shortcuts 1-103 Trim Start/End to Playhead Works in the Fairlight Timeline 1-104 New Fade In and Out to Playhead Trim Commands 1-104 Sync Offset Indicator 1-104 Chapter – 6 Fairlight Page Improvements 1-82 New Fairlight Page Features This section describes some of the biggest new features added to the Fairlight page in DaVinci Resolve 15. ADR (Automated Dialog Replacement) Clicking the ADR button on the Interface Toolbar opens up the celebrated Fairlight ADR panel, which provides a thoroughly professional workflow for doing automated dialog replacement. Dialog replacement, for those who don’t know, is the process whereby audio professionals bring in actors to re-record unsalvageably bad dialog recordings from the comfort of their recording studios, line by line and with a great deal of patience. The ADR panel on the Fairlight page It’s an old joke is that ADR isn’t really automatic, but the Fairlight page aims to give you all the help it can to make this a structured and straightforward process. Simple yet powerful cue list management lets you efficiently assemble a re-recording plan. Industry-standard audio beeps and visual cues via your BMD video output device help the actors in the booth nail their timings and their lines. Then, sophisticated take management with star ratings and layered take organization in the Timeline help you manage the resulting recordings to pick and choose the best parts of each take when you edit the results. The ADR Interface When open, the ADR interface consists of three panels to the left of the Timeline, a Record panel, a List panel, and a Setup panel. The controls of these panels are described in the order in which they’re used. The Setup Panel As its name implies, the Setup panel is where you configure your ADR session. Chapter – 6 Fairlight Page Improvements 1-83 The Setup panel of the ADR interface This panel presents the following controls: Pre Roll and Post Roll: Specifies how many seconds to play before and after each cue’s specified In and Out points, giving actors a chance to listen to what comes before and after each cue in order to prepare. If you enable the Beep options below, beeps provide a countdown during the specified pre-roll. Record Source: (Disabled until you select a Record Track) A pop-up menu lets you choose the input you want to record from, creating a patch to the Record Track. Record Track: A pop-up menu lets you choose the track you want to record to. Selecting a track with this menu creates a patch from the Record Source to the Record Track, and automatically toggles Record Enable on. Guide Track: A pop-up menu lets you choose which track the original production audio you need to re-record is on. This is used for sending audio playback to the talent to use as a guide for recording their own replacement performance. Record File Name: A text entry field that lets you provide a name for the audio files being recorded to be saved with. Character List: A list for adding the names of all the characters that have dialog cues you’ll be re-recording, to help with cue creation and management. An Add New button lets you add additional names to this list, while a Remove button lets you delete characters you no longer need. Beep to In Point: Enables a three-beep sequence to be heard leading up to the recording. Beep at In Point: Enables one last beep at the In point. Chapter – 6 Fairlight Page Improvements 1-84 Count In: An onscreen counter that counts down to the start of the cue. Video Streamer: A visual cue for the talent to watch during pre roll to ready them for recording. A pair of vertical lines superimposed over the program being output to video that move towards one another across your video output screen during the pre-roll to the cue to give the talent a visual indication of the countdown to the beginning of the cue. When the beeps play, these lines stretch upward and down briefly. Both lines come together at the cue point, and a large cross is shown as recording begins. Smart Timeline: When turned on, this option automatically moves the playhead to each cue as it’s selected, and zooms in to frame the duration of that cue in the Timeline. Mixing Control: Enables automated switching of audio playback, to independently control what the talent and the audio engineer hear at various stages of the ADR recording process. For example, with this enabled, the Guide track is not routed to the Control room while the engineer is reviewing a take. The List Panel This is where you create a list of cues you need to re-record, either from within the Fairlight page, or imported from a .csv file that someone provides you. The List panel of the ADR interface This panel presents the following controls: Cue editing controls: Displays the data for the currently selected cue (or a cue that was just created). In and Out timecode fields store the timeline In and Out points that were set when the cue was created, but can be manually edited for fine tuning. A Character pop-up menu lets you choose which character that line of dialog belongs to. A text entry field lets you enter the dialog cue that’s to be re-recorded, so you and the talent can both refer to it. New Cue button: Clicking this button adds a new cue to the list using whatever In and Out points have been set in the Timeline, and whatever character was last selected. Cue List: The list of all cues that have been entered or imported. The Cue list can be filtered using the Filter pop-up menu at the top-right of the ADR panel (next to the option menu). You can choose to show the cues for all characters, or for any selected combination of characters. You can also choose to hide all cues that are marked as done to experience the joy of this list shrinking more and more the closer you are to being finished. Chapter – 6 Fairlight Page Improvements 1-85 Additionally, the ADR interface option menu has three commands pertaining to the List panel: Import Cue List: Lets you import a properly formatted .csv file to create cues that have been prepared in a spreadsheet. Correct formatting for cue lists you want to import is no headers, one line per cue, with four individual columns for In timecode, Out timecode, Character Name, and Dialog. Export Cue List: Lets you export the contents of the cue list to a .csv file, for exchange or safe-keeping. Clear Cue List: Deletes all cues in the cue list. It’s recommended you export a copy of your cue list before eliminating it completely, in case you ever need to revisit a cue. The Record Panel This is where you actually run the ADR recording session you’ve set up, using the dialog cues you’ve put into the Cue List The Record panel of the ADR Interface This panel presents the following controls: Record and rehearse controls: Four transport controls and two buttons let you control recording during ADR sessions. These controls are only clickable when you’ve selected a cue from the Cue List to record. –– Rehearse: Runs the section of the Timeline specified by a cue without actually recording anything, giving the talent an opportunity to run through their dialog and practice their timing and delivery. Beeps and on-screen streamers are not played during a rehearsal. –– Play: Plays the currently selected take from the Take list (described below). If no take is selected, the most recently recorded one on top is played. –– Stop: Immediately stops rehearsal, playback, or recording. –– Record: Initiates recording of the cue to the specified audio track, with cue beeps and video streamer cues. –– Keep Playing: At the end of a take you may wish to keep playing, so the talent can hear the next section of of the track. Pressing the Keep Playing button at any time, even while recording, results in post roll being ignored and normal playback resuming after the cue’s out time. Chapter – 6 Fairlight Page Improvements 1-86 –– Keep Recording: At the end of a take you may wish to keep recording until you manually stop. Pressing the Keep Recording button at any time, even while recording, results in the Out point of the current cue being ignored and recording continuing until you stop it. Take List: The Take list shows every take you’ve recorded for the current cue, with take number, name, and a five-star rating that you can set to keep track of which takes worked and which didn’t. Earlier takes are at the bottom of this list, while recent takes are at the top (the same order in which the corresponding layered audio clips appear in the timeline track they’ve been recorded to). Cue List: The list of all cues that have been entered or imported. The Cue list can be filtered using the Filter pop-up menu at the top-right of the ADR panel (next to the option menu). You can choose to show the cues for all characters, or for any selected combination of characters. You can also choose to hide all cues that are marked as done to experience the joy of this list shrinking more and more the closer you are to being finished. Cue List Done column: A sixth column appears in the Record panel only, labeled Done. It contains check boxes for each cue that you can turn on to keep track of which cues you’ve successfully finished. Additionally, the ADR interface option menu has one command pertaining to the Record panel: Record Early In: Enables recording during pre-roll, in the event you’re working with talent that likes to start early. Setting Up To Do An ADR Session Setting up to record ADR is straightforward, but requires some steps. Patching tracks in preparation to record ADR: 1 In the Timeline, create two new audio tracks, one to record ADR to, and another to route the Oscillator through to play preview beeps. 2 Open Fairlight > Patch Input/Output to choose the Patch window 3 Choose Audio Inputs from the Source pop-up menu, and choose Track Inputs from the Destination pop-up menu, then patch the audio input for your recording microphone to the track you created to record onto in step 1. 4 Next, choose Osc from the Source pop-up menu, and patch the Beep output to the track you created to preview beeps in step 1. 5 Close the Patch Input/Output window. 6 To make sure you can hear the preview beeps, open the Mixer (if necessary), click the Input pop-up menu at the top of the channel strip that shows “Beeps,” and choose Path Settings. When the Path Settings window appears, turn on the Thru button in the Record Level controls, and close the path settings. Thru mode places a track into a “live” input mode, bypassing track playback in order to play the patched input. Thru mode is typically used for external sources that you want to bring into a mix. While a track is in THRU, there’s need to arm the Record button, the external source is always feeding to the mix. 7 If you’re recording ADR to your main timeline, Solo the Guide Track, the Record Track, and the Beep Preview track to focus only on the audio you’re re-recording. Now you’re ready to configure the Setup panel. Chapter – 6 Fairlight Page Improvements 1-87 Configuring the Setup panel: 1 Open the ADR interface, and then open the Setup panel. 2 Choose the Pre Roll and Post Roll you want to use, in seconds. A pre roll of at least 3 seconds is recommended to give the talent time to get ready. 3 From the Record Source pop-up menu, choose the microphone you patched earlier. 4 From the Record Track pop-up menu, choose the Record Track you created. 5 From the Guide Track pop-up menu, choose the track with the original production audio that you’re replacing. 6 At the bottom of this panel, turn on the Preroll Cue options you and the talent want to use. Options include Beeps to provide an audible lead up to re-recording, and a Video Streamer that gives both visual cues and also displays the dialog text for that cue on screen for the actor to refer to, keeping their eyes on the screen. Creating and Importing ADR Cue Lists You must have a list of cues to be able to use the ADR interface properly. There are two ways you can create a Cue list to record with, make one on the Fairlight page, or import one. If you’ve been doing all of your dialog editing inside of DaVinci Resolve, you can go ahead and create a list by marking the sections of the Timeline you need to re-record and creating cues from those timings. To manually add cues to the Cue List: 1 Open the Setup panel of the ADR interface, and use the Add New button in the Character Setup list to create all the character names you’ll be creating dialog cues for. These will help you to filter and sort the list as necessary later on. 2 Next, open the List panel of the ADR interface. This is where all the controls for creating and editing cues are. 3 In the Timeline, set In and Out points to mark the section of dialog you want to turn into a cue. Those timecode values appear in the Cue Editing section of the List panel. 4 In the Cue Editing section, choose the character who’s speaking that cue, and type the dialog they’re speaking. 5 When you’re done, click New Cue, and that cue will be added to the Cue List. 6 Repeat steps 3 through 5 until you’re finished creating all the cues you intend to re‑record. If you need to edit any cue, simply click to select that cue, and edit it in the Cue Editing section above. To import a .csv file to the Cue List: 1 Choose Import Cue List from the ADR option menu, then use the dialog to choose the .csv file containing the cue list you were given, and click Open. 2 An ADR Setup dialog appears letting you assign the columns of the .csv file you selected to the relevant column of the ADR panel. Correct formatting for Cue lists you want to import is no headers, one line per cue, with four individual columns for In timecode, Out timecode, Character Name, and Dialog, but if any of these columns are transposed, you can correct this here. Chapter – 6 Fairlight Page Improvements 1-88 Dialog for rearranging columns of cue data, if necessary 3 Click Import CSV. The cues should appear in the Cue List. To export a .csv file from the Cue List: Choose Export Cue List from the ADR option menu, choose a location to save the file, and click Save. Recording ADR to the Timeline Once you’ve configured your workstation for recording, and you’ve set up a cue list to work with, it’s time to start recording each cue. To record a cue from the Cue List: 1 Open the Record panel of the ADR interface. 2 If you want to record a particular character’s cues, you can turn off Show All Characters, and turn off all unnecessary characters, in the ADR option menu. 3 With the list showing the character cues you need, select the cue you want to start recording. That cue contains the timecode necessary to determine which part of the Timeline to record to, and the playhead automatically moves to that part of the Timeline. 4 Click the Rehearse button to run through the cue with the talent a few times. Both audio and video corresponding to that cue will play, including pre roll and post roll. 5 When the talent is ready to try a take, click the Record button, and let the Fairlight page do the work of playing through pre roll with beep notifications and visual streamer cues, initiating recording, and then stopping recording automatically once the cue is done. 6 If you or the talent want to hear what the take sounds like again, you can select it in the Take List and click Play. Depending on how you like the take, you can mark it with the 5-star ratings control and then record another take. 7 To record another take, click the Record button again. New takes are placed on the same track using layered audio, so you can record as many takes as you like into the same area of the Timeline that corresponds to the media you’re replacing. When you’re finished recording takes, you’ll have a neatly organized stack of alternate takes to draw upon as you edit the best parts from each recording. 8 When you’re finished recording a cue, click the Done checkbox for that cue, and select the next cue you want to record. When you’re finished re-recording dialog, simply close the ADR interface. Chapter – 6 Fairlight Page Improvements 1-89 Fixed Playhead Mode Choosing View > Show Fixed Playhead puts the Fairlight timeline into an audio-centric mode where the playhead remains fixed in place, and the Timeline scrolls underneath it as you use the Transport controls or JKL to play, shuttle, or scrub forward or back. Video and Audio Scrollers Checkboxes in the Timeline View Options let you optionally show one “Video Scroller” and up to two “Audio Scrollers” at the bottom of the Fairlight timeline. Video and Audio Scrollers at the bottom of the Fairlight timeline At the default “Low” zoom level, the Video Scroller provides a scrollable frame-by-frame filmstrip view of the video of your program, where one frame of the scroller equals one frame of your video. Each of the two Audio Scrollers, on the other hand, let you focus on a continuous waveform view of a particular audio track. You choose which track populates an Audio Scroller via a pop-up menu in the Timeline header. Audio Scrollers showing the pop-up menu that selects which track they display What Are They Used For? The Audio Scrollers always provide a zoomed-in view of specific audio tracks that you’re focused on, regardless of the zoom level of the Timeline tracks above. This means you can focus on subtle details of the audio of one or two tracks that you’re working on, while the rest of the Timeline shows you the overall stack of tracks with clips that are playing together at that moment. Meanwhile, the Video Scroller always shows the exact frame of video that corresponds to the current moment in time, so it’s an aid to precision editing involving frame-specific adjustments. Chapter – 6 Fairlight Page Improvements 1-90 Additionally, both the Filmstrip and Waveform viewers scroll continuously during playback, giving you a preview of what visual actions and audio cues are coming a few moments forward in time that you can refer to while performing automation or recording foley. Repositioning the Scroller Playhead While the Scrollers are visible, the Scroller playhead can be dragged to the left or right in the Timeline to give you more or less preview room to the right. Zooming the Video Scroller Right click on the Video scroller lets you choose a Low, Medium, or High zoom level. At Low, you get a frame-by-frame view of the program that feels like scrolling a strip of film on a Steenbeck. At Medium and High, you get a progressively abbreviated film strip that lets you zoom more quickly. Scrolling the Fairlight Timeline Using the Scroller Tracks Dragging the scroller tracks to the left or right smoothly scrubs through the Timeline in greater detail, regardless of the zoom level of the Timeline tracks above. 3D Audio Pan Window Option-double-clicking on the Pan control of the Mixer opens an alternate 3D Audio Pan window. Whereas the regular Pan window lets you do stereo and conventional 5.1 and 7.1 surround panning, the 3D Audio Pan window lets you do the kind of spatial audio positioning enabled by advanced surround formats such as Auro 3D and NHK 22.2 (more information about specific support for these and other formats will come later). The 3D Pan window The 3D Audio Pan window has a few more controls than the ordinary Pan window: Pan enable: Toggles the entire panning effect on and off. Panner viewer: A large 3D representation of the listener’s perceived sound stage, with a blue sphere that represents the position of the track’s audio being positioned within that space, that casts a shadow straight down on the floor and projects a blue box on the four walls of this space to indicate its position more concretely. Chapter – 6 Fairlight Page Improvements 1-91 Front panner: A 2D panning control that represents the horizontal Left/Right axis and the vertical Up/Down axis, letting you make these specific spatial adjustments. Side panner: A 2D panning control that represents the horizontal Front/Back axis and the vertical Up/Down axis, letting you make these specific spatial adjustments. Top panner: A 2D panning control that represents the horizontal Left/Right axis and the vertical Front/Back axis, letting you make these specific spatial adjustments. Left/Right: A 1D knob that changes the balance of signal between the left and right side speakers you’re outputting to, depending on what speaker format you’re mixing to. Front/Back: A 1D knob that changes the balance of signal between the front and back sets of speakers you’re outputting to, depending on what speaker format you’re mixing to. Rotate: A 1D knob that simultaneously adjusts the left/right and front/back pan controls in order to horizontally rotate a surround mix about the center of the room. Tilt: A 1D knob that simultaneously adjusts the left/right and Up/Down pan controls in order to vertically rotate a surround mix about the center of the room. Spread: Only available when a linked group is selected. Spread adjusts the perceived size of a surround mix. Divergence: Spreads the signal for an individual feed across more of the adjacent loudspeakers, making the perceived size of the sound source larger. Boom: The send level of that track to the LFE part of the mix. An On button enables this functionality, while a Pre button lets you adjust the “dry” part of the signal separately from the “Wet” part of the signal when effects are applied. Visible Video Tracks A checkbox in the Timeline View Options pop-up menu of the toolbar lets you display small versions of the video tracks in the Fairlight timeline, for reference. These video tracks are uneditable; they’re simply there so you can see which audio clips correspond to which video items, and so they can be used as snapping targets for positioning audio. Showing video tracks on the Fairlight page Chapter – 6 Fairlight Page Improvements 1-92 User-Selectable Input Monitoring Options The Fairlight > Input Monitor Style submenu presents five options governing how you want to monitor inputs while recording. Input: You only hear the live signal being input, you never hear the contents of tracks. Auto: When one or more tracks is armed for recording, you hear the live input signal, on playback you hear the contents of each track. Record: You only hear the live input signal while actively recording, meaning the Record button has been pressed while one or more tracks are armed for recording. You don’t hear the input signal while tracks are merely armed. Mute: You hear nothing. Repro: While recording, you only hear what’s just been recorded, played from the track. In other words, you’re not listening to the live input, but you’re reviewing what’s just been recorded as it’s recording. Commands For Bouncing Audio Two new commands are available for bouncing audio on the Fairlight page. Bouncing audio refers to mixing and rendering audio from Timeline tracks to another track on the Timeline, in the process “baking in” processor intensive effects and complicated fixes to create a new piece of audio. Timeline > Bounce Selected Tracks to New Layer Timeline > Bounce Mix to Track To use Bounce Selected Tracks to New Layer: 1 Set In and Out points to define the range of the Timeline you want to bounce. If you don’t do this, nothing will happen. 2 Command click the track header or mixer channel strips of the tracks you want to bounce to select them. 3 Choose Timeline > Bounce Selected Tracks to New Layer. The audio on each track is processed and rendered and appears as the top layer of audio on that track. The original audio with live effects is still available as the bottom of the stack of layered audio on that track. To use Bounce Mix to Track: 1 Choose Timeline > Bounce Mix to Track. The Bounce Mix to Track window appears, showing each Main, Submix, and Auxiliary that’s currently available. 2 In the Destination Track column, set which mixes you want to bounce by choosing either New Track or a specific track from the pop-up menus, then click OK. The specified Mix is processed, mixed, and bounced to the specified track as a new piece of audio. Chapter – 6 Fairlight Page Improvements 1-93 Sound Library Browser A Sound Library panel is available from the Interface toolbar for browsing sound effects libraries that you have available to you, on your system or on a SAN you’re connected to. It includes the capability of scanning specified file paths to catalogue available sound files and their metadata, storing this data within the currently selected project database (or another database that you select) to use when searching for the perfect sound effect within your library. Once you’ve catalogued your sound effects collection, it’s easy to search for sounds, preview what’s been found in the list, and edit the one you like best into the timeline. The Sound Library panel The Sound Library panel has the following controls: List display controls: The Sound Library title bar has controls for sorting the sound effects list, showing it in List or Icon view, and an Option menu with various other settings and commands. Search field: Enter a term into the search field to look for sound effects files using that metadata. A pop-up menu to the right lets you search the database by Name or Description metadata. Library controls: Clicking the Library button (to the right of the Search field) reveals a menu that lets you choose which database to use for searching (and cataloging) sound effects collections. Each PostgreSQL database can have a different catalog. Chapter – 6 Fairlight Page Improvements 1-94 Choosing a library to search Preview and Audition controls: These controls let you preview and audition sound effects that you find as you look for the right one. –– Clip name: The name of the current clip you’ve selected. –– Next/Previous buttons: Two buttons let you select the next or previous sound effect clip in the Sound Effect list. –– Zoom control: Controls the zoom level of the Playthrough waveform. –– Duration field: Shows the duration of the current clip, or of the section of the clip marked with In and Out points. –– Playhead timecode field: The timecode of the playhead’s position. –– Navigation waveform: The waveform of the entire sound effect appears here, making it easy to jump to different parts of the selected clip. All channels are summed together in this display. –– Playthrough waveform: A zoomed-in section of the selected clip that lets you see more waveform detail for setting In, Out, and Sync points. –– Jog Bar: Lets you scrub around the clip. –– Transport controls: Stop, play, and Loop buttons let you control playback, although you can also use the space bar and JKL controls. Right-click the Stop button to switch it into “Stop and Go to Last Position” mode. –– Marking controls: The sync point button lets you mark which frame of the sound effect you want to use to sync to a frame of the Timeline when you audition. In and Out points let you mark how much of the sound effect clip you want to edit into the Timeline. –– Audition controls: The Audition button puts you into Audition mode where the currently selected sound effect clip appears at the position of the playhead in the currently selected Timeline track. Cancel and Confirm buttons let you choose whether you want to remove the clip from the Timeline and try again with another clip, or leave the sound effect clip in. Sound Effect list: All sound effects clips that match the current search criteria appear in this scrollable list. Each item in the list shows its name, duration, channel mapping, and star ratings that you can customize. To catalogue all audio files within a given file path: 1 Open the Sound Library. 2 (Optional) Click the Library button (to the right of the Search field), and select which PostgreSQL-based project database you want to save the resulting metadata analysis to using the Library pop-up menu that appears. The current database is selected by default. If you’re using a disk-database instead, the top compatible PostgreSQL database in the list will be the default. 3 Click the Option menu and choose Add Library. From the file dialog that appears, select the top-most directory of a file path that contains sound effects; if you’ve selected a directory with sub-directories inside, each subdirectory will be examined for content. 4 Click Open. Chapter – 6 Fairlight Page Improvements 1-95 A progress bar will show you how long the operation will take. When you’re finished, a dialog will appear letting you know how many clips were added to the current library. To search for a specific sound effect and edit it into the Timeline: Type a search term into the Search field. Optionally, you can click the Library button to the right of the search field, and use the Type, Duration, and Format pop-up menus to help limit your search. All audio cues that include the search term in their file names will appear in a list below. Selecting an item on the list loads it into the preview player where you can play it or audition it in your timeline. Auditioning clips you’ve found in the Timeline: 1 Select a sound effect clip you’ve found from the list that you want to audition in the Timeline. 2 Move the playhead to the part of the sound effect that you want to sync to, and click the Sync Point button to place a sync mark on that clip. For example, if you’re syncing a car door closing, you might sync to the peak of the “slam.” 3 Set In and Out points to define the range of the sound effect you want to potentially use. 4 Select a track you want to preview the sound effect in by clicking its track header or Mixer channel strip. 5 Position the playhead at the place in the Timeline you want to align the sync mark in the sound library to. 6 Click the Audition button in the Sound Library. That clip now appears, temporarily, in the Timeline, and you can play through that section of the Timeline to see how you like the sound effect in context with the rest of the mix. 7 If you like the sound effect, click Conform to keep it in the Timeline. If you don’t, click Cancel. VSTi Support For Recording Instrument Output DaVinci Resolve 15 introduces support for VSTi instruments working with connected MIDI controllers to trigger instrument sounds that can be recorded live on audio tracks of the Timeline. This is intended to be used for loading a VSTi sampler with foley sounds such as footsteps or punches, so you can perform these sounds in real time and record the result to another track as you watch performers walking or punching in the edit, even if you lack a recording booth with foley pits and props. On the other hand, if you’re a musician, there’s nothing stopping you from loading VSTi musical instruments of different kinds for playback, and using the Fairlight page as a multi-track recorder. DaVinci Resolve doesn’t have MIDI sequencing functionality, but you can record live playback straight to the Timeline, using layered audio to manage multiple takes for later re-editing. Bet you never thought you’d be recording music in DaVinci Resolve…. Chapter – 6 Fairlight Page Improvements 1-96 A VST Instrument (in this case Serato Sample) loaded into a track of the timeline To enable a MIDI controller in macOS: 1 If DaVinci Resolve is running, quit before connecting your MIDI controller and setting it up. 2 On macOS you’ll use the Audio Midi Setup utility to choose output hardware and select a speaker configuration to be made available on your system. In the Finder, use Spotlight and search for Audio MIDI Setup to open it. 3 In Audio MIDI Setup, choose Window > Show MIDI Studio. A window showing icons for all connected MIDI controllers appears. Your controller should be showing an icon. If it’s not, you may need to install drivers for it. 4 Select the icon for your controller and turn on the “Enter test MIDI setup mode” button (it looks like a little keyboard) to test if your keyboard is connecting with the computer. If it is, then turn this off. For more information on setting up MIDI on different systems, see the DaVinci Resolve Configuration Guide, available on the web from the Blackmagic Design support page at https://www.blackmagicdesign.com/support/family/davinci-resolve-and-fusion. To set up the Fairlight page for VSTi instrument recording using a sampler: 1 Open DaVinci Resolve 15. 2 Make sure you have at least two available audio tracks in the Timeline, one for the instrument you’ll be playing, and one to record into. This example will use tracks A4 and A5 for this. 3 Open the Effects Library, find a VSTi sampler you have installed on your system, and drag it to the track header of the track you want to use for playing, for example track A4. Massively-featured sampler/synth combinations such as Native Instruments Kontakt and Steinberg Halion are ubiquitous and useful when you want to specifically map a collection of sound effects to specific keys or pads to create re-usable multi-purpose instruments. However, more streamlined samplers that emphasize automatic audio clip Chapter – 6 Fairlight Page Improvements 1-97 slicing such as Serato Sample (Windows and macOS) or Image Line Slicex (Windows only) can make short work of the more specialized task of loading library sound effects recordings (or custom recordings you create) with multiple footsteps, punches, keyboard presses, cloth rustles, or other foley activities, and quickly splitting them up into individually playable samples you can trigger with pads or a keyboard. 4 When the VSTi interface window appears, open the MIDI menu at the upper righthand corner of the VSTi window and choose the correct MIDI channel from your MIDI controller’s submenu. If you’ve selected the correct MIDI channel, the instrument should start responding to the keys or pads on your controller. Enabling MIDI control 5 Next, configure the VSTi instrument you’re using to play the sound effects you want to use for foley. In this example, the Serato Sample VSTi plug-in is being used to automatically slice up a recording of footsteps from one of Sound Ideas’ many sound effects libraries. Because the VSTi you added is patched to that track’s Insert (if you look at the Mixer you should see that the I button is enabled on the channel strip the instrument is patched to), the Send is PRE the Instrument. This means you need to patch that track’s Track Direct output to the input of another track to record the instrument. 6 Choose Fairlight > Patch Input/Output to open the Patch Input/Output window, then set the Source pop-up menu to Track Direct and the Destination pop-up menu to Track Input. Click Audio 4 to the left, and Audio 5 to the right, and click the Patch button; this sets you up to play the VSTi plug-in on track A4, and record its output on track A5. Be aware that after patching Track Direct from the track with the instrument to the track you’re recording onto, you also need to turn “Direct Output” on for that track in the Path Settings of that track’s channel strip in the Mixer. 7 Open the Mixer (if necessary), click the Input pop-up menu at the top of the channel strip that shows the VSTi instrument you’re using, and choose Path Settings. When the Path Settings window appears, click the ON button for Direct Output, then close the path settings window. At this point, you’re ready to begin recording. Chapter – 6 Fairlight Page Improvements 1-98 To play and record a VSTi instrument: 1 Click the Record Arming button of the track you’re recording to (in this example A5), move the playhead to where you want to begin recording, and then click the Record button to begin recording. 2 As the video of your program plays, use your MIDI controller to trigger sound effects as necessary. When you’re finished, click the Stop button. If necessary, you can record multiple takes using track layering until you get the timing right. When you’re finished, you can remove the instrument from the track it’s on since the recorded audio is all you need. General Fairlight Page Enhancements A wide variety of enhancements and new features have been implemented throughout the Fairlight page to improve everyday workflows. Normalize Audio Levels Command A Normalize Audio Levels command automatically adjusts the level of clips to peak at a specific target level, measured in dBFS. This is only a volume adjustment, no dynamics are applied, so the result of using this command is that the loudest parts of each selected clip are going to match one another at the target level. This command is also available in the Edit page. To normalize one or more selected audio clips: 1 Right-click one of the selected clips and choose Normalize Audio Levels 2 A dialog appears with two options. Choose the Reference Level that you want to set the peak volume of the selected clips to match, and then choose how you want to set the level of multiple selected clips: When Set Level is set to Relative, all selected clips are treated as if they’re one clip, so that the highest peak level of all selected clips is used to define the adjustment, and the volume of all selected clips is adjusted by the same amount. This is good if you have a series of clips, such as a dialog recording, where the levels are consistent with one another, and you want to normalize all of them together. When Set Level is set to Independent, the peak level of each clip is used to define the adjustment to that clip, so that the volume of every selected clip is adjusted by an amount specific to that clip. The end result may be a set of very different volume adjustments intended to make the peak levels of each audio clip match one another. This is good if, for example, you’re trying to balance a series of different sound effects with one another that have very different starting levels. Clip Pitch Control Selecting a clip and opening the Inspector reveals a new set of Clip Pitch controls that let you alter the pitch of a clip without changing the speed. Two sliders let you adjust clip pitch in Semi Tones (large adjustments, a twelfth of an octave) and Cents (fine adjustments, 100th of an octave). Clip Pitch Control in the Inspector Chapter – 6 Fairlight Page Improvements 1-99 Support for Mixed Audio Track Formats from Source Clips DaVinci Resolve 15 now supports media with multiple audio tracks that have differently formatted channels embedded within them. For example, a clip with one stereo track, one 5.1 surround track, and six mono tracks can all be appropriately set up in the Audio panel of Clip Attributes after that clip has been imported. The Audio panel of Clip Attributes now has controls over what format (Mono, Stereo, 5.1, 7.1, Adaptive) the channels embedded within a particular clip should be configured as. This means that you can set up clips with multiple tracks, each one using potentially different formats of audio employing different combinations of clips, which is handy for mastering. Clip Attributes now lets you assign channels among different tracks with different channel assignments Oscillator for Generating Tone, Noise, and Beeps The Fairlight page has a general purpose Oscillator, the settings of which you can customize by choosing Fairlight > Oscillator Settings. This opens the Oscillator Settings window that you can configure to generate tones, noise, or beeps using five sets of controls: Enable/Disable Oscillator toggle: Lets you turn the Oscillator on or off system-wide. Frequency dial: Sets a custom frequency of oscillating tone, from 20 Hz to 10kHz. Defaults to 1kHz. Level dial: Sets the output level for the tone or noise, from –50dB to +10dB. Defaults to –15 dB. Frequency presets: Four buttons let you choose from four commonly used tones, 100 Hz, 440 Hz, 1 kHz, and 2 kHz. Noise type buttons: Two buttons let you choose from White noise or Pink noise. Chapter – 6 Fairlight Page Improvements 1-100 You can set up the Oscillator to output whatever kind of tone or noise you require, and then patch it to tracks for recording tones, or patch it to audio outputs for calibrating speakers. If you use the beep options of the ADR panel, those are performed via the Oscillator. To play the Oscillator out of your speakers: 1 Choose Fairlight > Patch Input/Output to open the Patch Input/Output window. 2 Choose Osc from the Source pop-up menu, and choose Audio Outputs from the Destination pop-up menu. 3 At the left, click the button of what you want to output, Osc (Oscillator) or Noise. 4 At the right, click the connected audio outputs that you want to patch to, and click Patch. Tone or noise should immediately start playing out of your configured speakers. 5 To stop, select one of the patched buttons, and click Un-Patch. To record a tone or noise from the Oscillator to an audio track: 1 Choose Fairlight > Patch Input/Output to open the Patch Input/Output window. 2 Choose Osc from the Source pop-up menu, and choose Track Input from the Destination pop-up menu. 3 At the left, click the button of what you want to output, Osc (Oscillator) or Noise. 4 At the right, click the connected audio outputs that you want to patch to, and click Patch. Close the Patch Input/Output window. 5 Click the Arm Record (R) button in the track header of the track you patched the Oscillator to. If your Main is properly patched to your outputs, you should hear the tone or noise, and that track’s audio meter should immediately rise to the level being output by the Oscillator. 6 Click the Record button of the Transport controls to initiate recording of that tone to the patched track. Click the Stop button or press the spacebar to halt recording when you’re done. Compound Clips Breadcrumb Controls Below Fairlight Timeline When compound clips containing audio are opened in the Fairlight page, breadcrumbs controls appear beneath the Timeline that let you exit the compound clip and get back to the master Timeline. Level and Effects Appear in Inspector for Selected Bus When you select a Main, Submix, or Auxiliary bus channel strip in the Mixer, the Inspector updates with the Volume and any plug-ins that have been applied to that bus. Audio Waveform Display While Recording DaVinci Resolve is able to draw an audio waveform for audio that’s being recorded, in real time. This gives you immediate feedback that the input you’re recording is properly connected or not. Audio Playback for Variable Speed Clips Video/Audio clips with variable speed effects applied to them can now play either pitch‑corrected or un-pitch-corrected variable speed audio. An option in the Speed menu of the Retime controls lets you choose whether or not the audio is pitch-corrected. Chapter – 6 Fairlight Page Improvements 1-101 IMPORTANT: This feature works great with clips that were imported linked into the Media Pool, or that were synced in the Media Pool and then edited into the Timeline linked. However, clips that were manually linked together in the Timeline using the Link Clips command won’t play audio properly if you create variable speed effects. However, there’s a quick fix, which is to drag manually linked clips from the Timeline to the Media Pool to create a new source clip, and then drag that new source clip back to the Timeline. Paste and Remove Attributes for Clips, Audio Tracks The Fairlight page now has Paste Attributes and Remove Attributes commands that allow for the copying and resetting of audio parameters and effects, similar to the same commands on the Edit page. Loop Jog Scrubbing Currently available only on the Fairlight page, choosing Timeline > Loop Jog enables a brief sample preview to be heard while scrubbing the playhead through the Timeline. This can make it easier to recognize bits of dialog or music as you’re quickly scrubbing through tracks, in situations where you’re trying to locate specific lines or music cues. It also enables this brief sample preview to loop endlessly when you hold the playhead on a frame, so you can pause while scrubbing and hear (by default) the current 80 ms prior to the playhead as it loops. A pair of settings in the User Preferences let you customize this behavior. Loop Jog Alignment: Three options let you choose whether you loop audio Pre the position of the playhead, Centered on the playhead, or Post the position of the playhead. Loop Jog Width: A field lets you choose how many milliseconds of audio to loop when Loop Jog is enabled. How many milliseconds of audio corresponds to one frame depends on the frame rate of the video. For example, at a frame rate of 25 fps, there are 1000/25 = 40 ms per frame, so the default value of 80 ms equals two frames of looping. Improved Speaker Selection Includes Multiple I/O Devices The Speaker Setup controls now provide the option to assign specific Monitor Sets to specific audio I/O devices via a new Device pop-up menu, making it possible to listen to different speakers via different audio I/O boxes. Every compatible audio I/O device connected to your workstation appears in the Device pop-up menu. Assigning different audio I/O devices to different speaker configurations Chapter – 6 Fairlight Page Improvements 1-102 Fairlight Page Editing Enhancements The following edit-specific enhancements have been added to the Fairlight page. Media Pool Preview Player The Media Pool has a preview player at the top that provides a place to open selected source clips in the Media Pool, play them, add marks to log them, and set In and Out points in preparation for editing them into the Timeline via drag and drop. The Media Pool Preview Player effectively acts as a Source monitor for editing in the Fairlight page. The preview player in the Media Pool Various viewing controls populate the title bar at the top. A pop-up menu at the upper left lets you choose a zoom level for the audio waveform that’s displayed. To the right of that, a Timecode window shows you the duration of the clip or the duration that’s marked with In and Out points. Next to the right, a real-time performance indicator shows you playback performance. In the center the title of the currently selected clip is shown, with a pop-up menu to the right that shows you the most recent 10 clips you’ve browsed. To the far left, a Timecode field shows you the current position of the playhead (right-clicking this opens a contextual menu with options to change the timecode that’s displayed, and to copy and paste timecode). The center of the Media Pool Preview Player shows you the waveforms in all channels of the currently selected clip, at whatever zoom level is currently selected. Transport controls at the bottom consist of a Jog bar for scrubbing, Stop, Play, and Loop buttons, and In and Out buttons. Edit Page—Compatible Navigation and Selection Keyboard Shortcuts Certain navigation and selection keyboard shortcut have been changed in the Fairlight page to be consistent with identical operations on the Edit page: Previous Frame (Left Arrow) Next Frame (Right Arrow) Previous Edit (Up Arrow) Next Edit (Down Arrow) Timeline > Audio Track Destination Chapter – 6 Fairlight Page Improvements 1-103 The previous Fairlight commands that were mapped to the Arrow keys have been preserved, but now require modifiers to be used: Up and Down Arrow is now Command-Option-Up/Down Arrow. Left and Right Arrow is now Command-Option-Left/Right Arrow. Trim Start/End to Playhead Works in the Fairlight Timeline The following Edit page trim commands are now available for editing on the Fairlight page: Trim Start (Shift-Left Bracket) Trim End (Shift-Right Bracket) New Fade In and Out to Playhead Trim Commands A pair of commands in the Trim menu let you move the playhead over a clip, and use the playhead position to “Fade In to Playhead” or a “Fade Out to Playhead.” These commands work in both the Edit and Fairlight pages. (Left) Placing the playhead where you want a fade in to end, (Right) Using Fade In to Playhead Sync Offset Indicator Audio clips in the Fairlight page now display “out-of-sync” or sync offset indicators when they’re moved out of sync with the video items they’re linked to. If you’ve moved an audio clip out of sync with the video clip it’s linked to, there’s an easy way of getting them back into sync, by right-clicking the red out-of-sync indicator of any clip and choosing one of the available commands: Slip into place: Slips the content of the selected clip, without moving the clip, so that it’s in sync with the other items that are linked to that clip. Move into place: Moves the selected clip so that it’s in sync with the other items that are linked to that clip. Chapter – 6 Fairlight Page Improvements 1-104 Chapter 7 FairlightFX DaVinci Resolve 15 introduces FairlightFX, a DaVinci Resolve-specific audio plug-in format that runs natively on macOS, Windows, and Linux, providing high-quality audio effects with professional features to all DaVinci Resolve users on all platforms. Thirteen new audio plug-ins that can be used both in the Edit and the Fairlight page include a wide variety of plug-ins for repairing faulty audio, creating effects, and simulating spaces. This chapter explains what they do and how to use them. Chapter – 7 FairlightFX 1-105 Contents Common Controls For All FairlightFX 1-107 Chorus 1-107 De-Esser 1-109 De-Hummer 1-110 Delay 1-111 Distortion 1-112 Echo 1-113 Flanger 1-114 Modulation 1-115 Noise Reduction 1-117 Pitch 1-119 Reverb 1-119 Stereo Width 1-121 Vocal Channel 1-122 Chapter – 7 FairlightFX 1-106 Common Controls For All FairlightFX Before going into the specific controls of each FairlightFX plug-in, there are some common controls that all plug-ins share, found at the top of the custom GUI window for each plug-in. Common controls for all FairlightFX Presets: A cluster of controls that let you recall and save presets specific to each plug‑in. –– Add Preset button: Click this button to save the current settings of the FairlightFX you’re using. A dialog lets you enter a Preset name and click OK. –– Preset pop-up menu: All presets for the currently open plug-in appear in this menu. –– Previous/Next preset buttons: These buttons let you browse presets one by one, going up and down the list as you evaluate their effects. A/B Comparison: A set of buttons that lets you compare two differently adjusted versions of the same plug-in. The A and B buttons let you create two sets of adjustments for that plug-in, and toggle back and forth to hear which one you like better. The arrow button lets you copy the adjustments from one of these buttons to the other, to save the version you like best while experimenting further. Reset: A single reset control brings all parameters in the current plug-in to their default settings. Chorus An effects plug-in. A classic Chorus effect, used to layer voices or sounds against modulated versions of themselves to add harmonic interest in different ways. An animated graph shows the results of adjusting the Modulation parameters of this plug-in, giving you a visualization of the kind of warble or tremolo that will be added to the signal as you make adjustments. The Chorus FairlightFX Chapter – 7 FairlightFX 1-107 Chorus has the following controls: Bypass: Toggles this plug-in on and off. Input Format: (Only visible when Echo is inserted on a multi-channel track.) Lets you choose how multiple channels are input to the echo. Stereo sets separate Left and Right channels. Mono sums Left and Right to both channels. Left inputs the Left channel only, and Right inputs the Right channel only. Delay: The amount of delay between the original sound and the Chorus effect. –– Delay Time: Length of the Chorus delay lines. –– Separation: Time separation of the delay voices. –– Expansion: Sets L/R length differences, phase offset of modulators Modulation: These controls adjust the low frequency oscillator (LFO) that drives the tremolo of the chorus effect in different ways. –– Waveform: Specifies the shape of the LFO that modulates the rate of the Chorus, affecting the timing of the oscillations. There are six options: Sine (smooth oscillations), Triangle (sudden oscillations), Saw1, Saw2 ( jerky oscillations), Square (hard stops between oscillations), and Random (randomly variable oscillations). –– Frequency: Rate of LFO controlling the Chorus. Lower values generate a warble, higher values create a tremolo. –– Pitch: Amount of frequency modulation, which affects the pitch of the Chorus. –– Level: Depth of level modulation. Affects the “length” of the segment of Chorus that’s added to the sound. Low values add only the very beginning of the Chorus effect, high values add more fully developed Chorus warble or tremolo. Feedback –– Amount (%): The percentage of signal fed back to the Chorus Delay Line. Values can be positive or negative, the default is 0 (no effect). Increasing this parameter adds more of the Chorus effect to the signal, lowering this parameter adds more of the inverted Chorus effect to the signal. At values closer to 0, only a faint bit of Chorus can be heard in the audio, but at values farther away from 0 (maxing at +/- 99), a gradually pronounced Chorus becomes audible. –– Bleed (Hz): Amount of feedback which bleeds into the opposite channel (Stereo mode only). Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. Chapter – 7 FairlightFX 1-108 De-Esser A repair plug-in specific to dialog. The De-Esser is a specialized filter that’s designed to reduce excessive sibilance, such as hissing “s” sounds or sharp “ts” sounds, in dialog or vocals. A graph shows you which part of the signal the controls are set up to adjust, while reduction and output meters let you see which part of the signal is affected and what level is being output. The De-Esser FairlightFX The De-Esser has the following controls: Bypass: Toggles this plug-in on and off. Frequency Range: Two controls let you target the frequency of the “s” sound for a particular speaker. –– Target Frequency: A knob that lets you target the frequency of the offending sibilance. Sibilant sounds are usually found in the range of 5 - 8kHz. –– Range: switches the operational mode of the De-esser. Three choices (from top to bottom) let you switch among Narrow Band, Wide Band, and All High Frequency which processes all audio above the source frequency. Amount: Adjusts the amount de-essing that’s applied. Reaction Time: Adjusts how suddenly de-essing is applied. There are three choices. –– Relaxed: Equivalent to a slow attack. –– Fast: Equivalent to a fast attack. –– Pre-emptive: A “lookahead” mode. Chapter – 7 FairlightFX 1-109 De-Hummer A repair plug-in with general applications to any recording. Eliminates hum noise that often stems from electrical interference with audio equipment due to improper cabling or grounding. Typically 50 or 60 cycle hum is a harmonic noise, consisting of a fundamental frequency and subsequent partial harmonics starting at twice this fundamental frequency. A graph lets you see the frequency and harmonics being targeted as you adjust this plug‑in’s controls. The De-Hummer FairlightFX De-Hummer has the following controls: Bypass: Toggles this plug-in on and off. Frequency: Target source fundamental frequency. A knob lets you make a variable frequency selection, while radio buttons let you select common frequencies that correspond to 50Hz/60Hz electrical mains that are the typical culprits for causing hum. Amount: Adjusts how much De-Hum extraction you want to apply. Slope: Adjusts the ratio of fundamental frequency to partial harmonics, the adjustment of which lets makes it possible for various kinds of hum to be targeted. For example, a value of 0 biases hum extraction towards the fundamental frequency, while a value of 0.5 gives equal extraction of all harmonics (up to 4), and finally a value of 1.0 targets the higher frequency partials. Chapter – 7 FairlightFX 1-110 Delay An effects plug-in. A general purpose stereo delay effect, suitable for tasks varying from track doubling, to early reflection generation, through simple harmonic enhancement. Processes in stereo or mono, depending on the track it’s applied to. A graph shows the timing and intensity of the echoes generated by this plug-in on each channel, and an Output meter displays the output level of the resulting signal. The Delay FairlightFX Delay has the following controls: Bypass: Toggles this plug-in on and off. Input mode: (Only visible when Delay is inserted on a multi-channel track.) Lets you choose how multiple channels are input to the delay. Stereo sets separate Left and Right channels. Mono sums Left and Right to both channels. Left inputs the Left channel only, and Right inputs the Right channel only. Filters: Alters the proportion of frequencies that are included in the delay effect. When the Delay plug-in is inserted on a Mono Channel, the Left and Right sections are replaced with a single “Delay” section. –– Low Cut (Hz): A global High Pass filter. –– High Cut (Hz): A global Low Pass filter. Delay: Adjusts the timing of the delay. –– Left/Right Delay (ms): Delay time of each channel. –– Left/Right Feedback (%): Feedback % of the Left or Right channel back to itself. A negative value equates to % of feedback with a phase reverse from the original signal. Feedback: Controls for adjusting the amount of bleed between channels. –– High Ratio: Adjusts the frequency of a damping filter for the feedback signal. –– Stereo Bleed: Adjusts the proportion of signal from Left and Right channel feedback which feeds into the opposite channel. When the Delay plug-in is inserted on a Mono channel, Stereo Bleed control does not appear. Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. Chapter – 7 FairlightFX 1-111 Distortion An effects plug-in. Creates audio distortion that’s useful for sound design and effects, ranging from simple harmonic distortion simulating an audio signal going through primitive or faulty electronics (such as bad speakers, old telephones, or obsolete recording technologies), all the way to mimicking an overdriven signal experiencing different intensities of hard clipping (think someone yelling through a cheap bullhorn, megaphone, or PA system). This plug-in includes soft tube emulation in the output stage. An animated graph shows the results of adjusting the Distortion parameters of this plug-in, giving you a visualization of the kind of harmonic distortion, waveshaping, and clipping that will be modifying the signal as you make adjustments. Input and Output meters let you see how the levels are being affected. The Distortion FairlightFX Distortion has the following controls: Bypass: Toggles this plug-in on and off. Filters: Two filters let you simulate devices reproducing limited frequency ranges. –– LF Cut: Low frequency distortion shaping. –– HF Cut: High frequency distortion shaping. Distortion: Three sets of controls let you create the type and intensity of distortion you want. –– Mode buttons: Switch the operational mode of distortion. The one to the left, Distortion, creates harmonic distortion. The button to the right, Destroy, is a more extreme polynomial waveshaper. –– Distortion: Adjusts the amount of distortion that’s applied to the signal. Higher values distort more. –– Ceiling: Adjusts the level of input signal that triggers clipping. Chapter – 7 FairlightFX 1-112 Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. –– Auto Level button: Applies automatic compensation for gain added to the signal due to the distortion being applied. Having this button turned on prevents the signal from becoming dramatically and unexpectedly increased, while turning it off frees you to do what you want, if what you want is to hear a lot of distortion. Echo An effects plug-in. A classic Echo effect, simulating the fate of the cursed Oread from Greek mythology. Processes in stereo or mono, depending on the track it’s applied to. A graph shows the timing and intensity of the echoes generated by this plug-in on each channel, and an Output meter displays the output level of the resulting signal. The Echo FairlightFX Echo has the following controls: Bypass: Toggles this plug-in on and off. Input Format: (Only visible when Echo is inserted on a multi-channel track.) Lets you choose how multiple channels are input to the echo. Stereo sets separate Left and Right channels. Mono sums Left and Right to both channels. Left inputs the Left channel only, and Right inputs the Right channel only. Filters: Alters the proportion of frequencies that are included in the delay effect. When the Delay plug-in is inserted on a Mono Channel, the Left and Right sections are replaced with a single “Delay” section. –– Low Cut (Hz): A global High Pass filter. –– High Cut (Hz): A global Low Pass filter. –– Feedback: Adjusts the frequency of a damping filter for the feedback signal. Chapter – 7 FairlightFX 1-113 Left Channel: Parameters that independently affect delay on the Left Channel. When the Echo plug-in is inserted on a Mono Channel, the Left Channel and Right Channel sections are replaced with a single “Echo” section with only the Delay Time, Feedback Delay, and Feedback controls. –– Delay Time: Global Delay time for the Left Channel. –– Feedback Delay: Echo Delay time for the Left Channel. –– Feedback: Feedback percentage of the Left channel back to itself. –– L > R Feedback: Percentage of Left feedback signal which feeds back to Right Channel. Right Channel: Parameters that independently affect delay on the Right Channel. –– Delay Time: Global Delay time for the Right Channel. –– Feedback Delay: Echo Delay time for the Right Channel. –– Feedback: Feedback percentage of the Right channel back to itself. –– R > L Feedback: Percentage of Right feedback signal which feeds back to Left Channel. Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. Flanger An effect plug-in, giving that unmistakable Flanger sound dating from the days of dual tape machines with a slight delay added to one in periodic intervals causing flanging as they got back in sync with one another. Typically used to add a sort of warbling harmonic interest to a signal, in a wide variety of ways. An animated graph shows the results of adjusting the Modulation parameters of this plug-in, giving you a visualization of the kind of warble that will be added to the signal as you make adjustments. The Flanger FairlightFX Chapter – 7 FairlightFX 1-114 The Flanger has the following controls: Bypass: Toggles this plug-in on and off. Input mode: (Only visible when the Flanger is inserted on a multi-channel track.) Lets you choose how multiple channels are input to the Flanger. Stereo sets separate Left and Right channels. Mono sums Left and Right to both channels. Left inputs the Left channel only, and Right inputs the Right channel only. Modulation: A low frequency oscillator (LFO) used to drive the Flanger effect. –– Waveform (Hz): Specifies the shape of the LFO that modulates the rate of the Flanger. The three choices are Sine (a smooth change in rate), Triangle (a jerky change in rate), and Sawtooth (an abrupt change in rage). Affects the timing of the warble that is added to the sound. –– Rate (%): Speed of the LFO, affects the speed of the warble that is added to the sound. Low rate values create a slow warble, while high rate values create more of a buzz. –– Depth: Affects the “length” of the warble that is added to the sound. Low values add only the very beginning of a warble, high values add more fully developed warble. Width: Consists of a single parameter, Expansion, which sets Left/Right channel length differences, along with the phase offset of modulators. Feedback: These controls determine, in large part, how extreme the Flanging effect will be. –– Amount (%): The percentage of signal fed back to the Delay Line. Values can be positive or negative, the default is 0 (no effect). Increasing this parameter adds more of the Flange effect to the signal, lowering this parameter adds more of the inverted Flange effect to the signal. At values closer to 0, only a faint phase shift can be heard in the audio, but at values farther away from 0 (maxing at +/- 99), a gradually increasing warble becomes audible. The type of warble depends on the Modulation controls. –– LPF Filter (Hz): Lets you filter the range of frequencies that will affect the feedback signal. Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. Modulation An effect plug-in. General purpose modulation plug-in for sound fx/design. Four effects combine an LFO, FM adjustment, AM adjustment, Sweep and Gain filters to allow simultaneous frequency, amplitude and space modulation. In conjunction with the additional Rotation controls, simple Tremelo and Vibrato effects can be combined with auto-filter and auto‑Pan tools in order to provide texture and movement to a sound. An animated graph shows the results of adjusting the Modulator, Frequency, and Amplitude parameters of this plugin, giving you a visualization of the kind of modulations that will be applied to the signal as you make adjustments. Output meters let you see what level is being output. Chapter – 7 FairlightFX 1-115 The Modulation FairlightFX Modulation has the following controls: Bypass: Toggles this plug-in on and off. Modulator: A low frequency oscillator (LFO), shown in blue in the animated graph. –– Shape: Specifies the shape of the LFO waveform that modulates the audio. Six options include Sine, Triangle, Saw1, Saw2, Square, Random. –– Rate (Hz): Adjust the speed of the modulating LFO. Lower settings result in warbling audio, while extremely high settings result in buzzing audio the timbre of which is dictated by the shape you’ve selected. Frequency: Frequency modulation (FM) of a secondary oscillator, shown as green in the animated graph. –– Level (%): Controls the amount of Frequency Modulation that’s applied, intensifying or easing off the effect. –– Phase: Since each of the four primary effects within this plug-in can be applied together, along with the fact that modulation with level components (Tremelo/ Rotation/Filter) have the ability to combine or cancel out one another, phase controls are available. Altering the phase of an individual effect allows control of such interaction (e.g., cancel out a high level change, or offset a cancellation). Filter: Sweep and gain filters. –– Level: Lets you set the amount of filter sweep and gain to additionally use to modify the signal. The amount you’ve selected is previewed in a 1D graph to the side. –– Tone: Adjusts the center frequency of sweep. –– Phase: Since each of the four primary effects within this plug-in can be applied together, along with the fact that modulation with level components (Tremelo/ Rotation/Filter) have the ability to combine or cancel out one another, phase controls are available. Altering the phase of an individual effect allows control of such interaction (e.g., cancel out a high level change, or offset a cancellation). Chapter – 7 FairlightFX 1-116 Amplitude: Amplitude modulation (AM) of a secondary oscillator, shown as green in the animated graph. –– Level: Amount of Amplitude modulation applied. (Disabled in Ring Modulation Mode.) –– Phase: Since each of the four primary effects within this plug-in can be applied together, along with the fact that modulation with level components (Tremelo/ Rotation/Filter) have the ability to combine or cancel out one another, phase controls are available. Altering the phase of an individual effect allows control of such interaction (e.g., cancel out a high level change, or offset a cancellation). –– Ring Modulation Mode: Enables a Ring Modulation effect (where the signal is multiplied by the modulator, rather than modulated by it). Rotation: These controls are only available when applied to a multi-channel track. –– Depth: Amount of Rotation applied. –– Phase: Since each of the four primary effects within this plug-in can be applied together, along with the fact that modulation with level components (Tremelo/ Rotation/Filter) have the ability to combine or cancel out one another, phase controls are available. Altering the phase of an individual effect allows control of such interaction (e.g., cancel out a high level change, or offset a cancellation). –– Offset: Start offset of rotation in order to further place the signal in space. Output: Controls for adjusting the final output from this plug-in. –– Dry/Wet (%): A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Output Level (dB): Adjusts the overall output level of the affected sound. Noise Reduction A repair plug-in designed to reduce a wide variety of noise in all kinds of recordings. Based on spectral subtraction, it’s able to automatically detect noise in sections of dialog, or it can be used manually by “learning” a section of noise that can be subsequently extracted from the signal. A graph shows a spectral analysis of the audio being targeted, along with a purple overlay that shows what noise is being targeted. Two audio meters let you evaluate the input level (to the left) versus the output level (to the right), to compare how much signal is being lost to noise reduction. The Noise Reduction FairlightFX in action Chapter – 7 FairlightFX 1-117 Noise Reduction has the following controls: Bypass: Toggles this plug-in on and off. Auto Speech Mode/Manual radio buttons: These buttons toggle the overall functionality of the Noise Reduction plug-in between two modes: Auto Speech Mode: Specifically designed for human speech/dialog, applying dialog extraction to the incoming signal to dynamically detect the noise profile outside of the detected speech. As a result, Auto Speech Mode does not require an initial a “learn” pass, and adapts itself better to noise that changes over time. Manual Mode: Enables the Learn button, as this mode requires the user to locate a section of the audio recording that is only noise that the plug-in can analyze. To initiate this analysis, position the playhead at the beginning of a section of the recording that is only noise, click the Learn button so it’s highlighted, and play forward through the noise, stopping before any sound you want to preserve is reached, and clicking the Learn button to turn it back off. A noise profile is generated (shown in Purple on the graph), which is subsequently extracted from the remaining signal. Threshold (in dB): Relates to the signal-to-noise ratio (SNR) in the source recording. Recordings with a poor signal-to-noise ratio will require a higher threshold value, resulting in more noise reduction being applied. Attack (in ms): Primarily useful in Auto Speech mode, this controls the duration over which the noise profile is detected. Ideally, the attack time should match the variability of the unwanted noise. A low value corresponds to a faster update rate of the noise profile and is useful for quickly varying noise; a high value corresponds to a slower update rate and can be used for noise that’s more consistent. Sensitivity: Higher sensitivity values exaggerate the detected noise profile; the result is that more noise will be removed, but more of the dialog you want to keep may be affected. Ratio: Controls the attack time of the signal profile relative to the attack time of the noise profile. A faster ratio detects and preserves transients in speech more easily, but the resulting speech profile is less accurate. Frequency Smoothing: Smooths the resulting signal in the frequency domain to compensate for harmonic ringing in the signal after the noise has been extracted. Time Smoothing: A toggle button enables smoothing of the resulting signal in the time domain as well. Dry/Wet: A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. Makeup Gain: To let you compensate for level that may be lost due to the noise reduction operation you’re applying, this applies a pre-gain in, from -6dB to +18dB, just before the dry/processed mix. Chapter – 7 FairlightFX 1-118 Pitch An effects plug-in. Shifts audio pitch without altering clip speed. The Pitch FairlighFX Pitch has the following controls: Bypass: Toggles this plug-in on and off. Semitones: A “coarse” adjustment that can shift audio pitch up to +/- 12 semitones. Cents: A “fine” adjustment that can tune audio pitch in +/- 100ths of a semitone. Dry/Wet: A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. Reverb A spatial simulation plug-in, capable of recreating multichannel reverberation corresponding to rooms of different sizes, adjustable via a graphical 3D cube control. This plug-in lets you take a “dry” recording and make it sound as if it’s within a grand cathedral, an empty room, or a tiled bathroom. To understand this plug-in’s controls, it helps to know that the signal follows three paths which are combined to create the final effect: A direct path. An early reflection path (ER) simulating early reflection rays obtained from the first multiple reflections on the walls, traveling from the virtual source to the virtual listener. A late reverberation path (Reverb) simulating the behavior of an acoustic model of the room. A graph shows an approximate visualization of the reverb’s effect on the frequencies of the audio signal. Chapter – 7 FairlightFX 1-119 The Reverb FairlightFX Reverb has the following controls: Bypass: Toggles this plug-in on and off. Room Dimensions: By controlling the size of the virtual room a sound is to inhabit, these parameters simultaneously control the configuration of Early Reflection and Late Reverberation processing. The acoustic modes from this simulated room are computed and fed to Late Reverberation processing. The shape, gain, and delay of the first reflections are computed and then fed to Early Reflection processing. –– Height, Length, Width: Defines the dimensions of the reverberant space, in meters. –– Room Size: The calculated Room Width x Length, in meters2. Reverb: Additional controls that further customize the configuration of Early Reflection and Late Reverberation processing. –– Pre Delay: Increase or negate the propagation time from the virtual source to the virtual listener. As a result, it modifies the initial delay time between the source signal and the first reflection. –– Reverb Time: Decay time of the Reverb tail. It controls the overall decay time of the acoustic modes from late reverberation processing. –– Distance: Modifies the distance between the virtual source and the virtual listener. It modifies only the configuration of early reflections processing. –– Brightness: Modulate the shape of the decay time over frequency. At maximum brightness, decay time is identical at any frequency. At minimum brightness, higher frequencies result in shorter decay time and therefore duller sound. –– Modulation: Adds random low-frequency phase modulation from the tapping point of ER processing. At 0%, modulation is not used. Early Reflection Tone: Four post equalization controls modify the tone of early reflections to suit a particular room’s characteristics. –– Low Gain –– Low Frequency –– High Gain –– High Frequency Chapter – 7 FairlightFX 1-120 Reverb Tone: Four post equalization controls modify the tone of the reverb tail to suit a particular room’s characteristics. –– Reverb Tail Low Gain –– Reverb Tail Low Frequency –– Reverb Tail High Gain –– Reverb Tail High Frequency Output: These controls recombine the three audio processing paths into a single output signal. –– Dry/Wet: A percentage control of the output mix of “dry” or original signal to “wet” or processed signal. 0 is completely dry, 100% is completely wet. –– Direct Level: The amount of the direct level to mix into the final signal. –– Early Reflection Level: The amount of early reflection to mix into the final signal. –– Reverb Level: The amount of reverb to mix into the final signal. Stereo Width An enhancement plug-in that increases or reduces the spread of a stereo signal in order to widen or reduce the separation between channels. If this plug-in is added to a Mono channel, it will be disabled, as there is no stereo width to either distribute or control. A graph shows the currently selected width of stereo distribution as a purple arc, while inside of that graph a stereo meter shows the Left and Right distribution of the audio signal. Two audio meters measure levels, an Input meter to the left, and an Output meter to the right. The Stereo Width FairlightFX in action Stereo Width has the following controls: Width: Lets you control the spread of the stereo output. Settings range from 0 (Mono) to 1 (Stereo) to 2 (extra wide stereo). Diffusion: Adds more complexity to the output. Sparkle: Adds more high frequencies to the spread. Chapter – 7 FairlightFX 1-121 Vocal Channel An enhancement plug-in for general purpose vocal processing consisting of Hugh Pass filtering, EQ, and Compressor controls. Side by side EQ and Dynamics graphs are presented above the controls. An output audio meter lets you monitor the final signal being produced by this plug-in. The Vocal Channel FairlightFX Vocal Channel has the following controls: High Pass: Enabled by a toggle, off by default. Has a single frequency knob that sets the threshold below which frequencies are attenuated to reduce boominess or rumble. EQ: A three-band EQ for fine tuning the various frequencies of speech, enabled by a toggle, including Low, Mid, and High Mode, Frequency, and Gain controls –– Low/Mid/Hi Mode: Lets you choose from different filtering options to use for isolating a range of frequencies to adjust. Different bands present different options. –– Low/Mid/Hi Freq (Hz): Lets you choose the center frequency to adjust. –– Low/Mid/Hi Gain (dB): Lets you boost or attenuate the selected frequencies. Compressor –– Thres (dB): Sets the signal level below which compression occurs. Defaults to -25dB. The range is from -40 to 0dB. –– Reaction: Adjusts how quickly compression is applied when a signal exceeds the threshold. The default is 0.10. –– Ratio: Adjusts the compression ratio. This sets the gain reduction ratio (input to output) applied to signals which rise above the threshold level. The default is 1.5:1. The range is 1.1 to 7.0. –– Gain (dB): Lets you adjust the output gain to compensate for signal lost during compression, if necessary. Chapter – 7 FairlightFX 1-122 PART 2 Fusion Page Manual Preview Chapter 8 Introduction to Compositing in Fusion This section of the DaVinci Resolve 15 New Features Guide, available for the public beta of DaVinci Resolve 15, is designed to give you a preview of the revised documentation that will accompany the eventual final release. As the initial version of the Fusion page in this public beta is very much an early release that will grow more feature complete over time, so too the new Fusion documentation is a work in progress that’s being substantially revised. The included sections have been designed specifically to help users who are new to Fusion get started learning this exceptionally powerful environment for doing visual effects and motion graphics, now available right from within DaVinci Resolve. Chapter – 8 Introduction to Compositing in Fusion 2-2 Contents DaVinci Resolve, Now With Fusion Inside 2-4 How Do I Use the Fusion Page? 2-4 How Do Fusion Effects Differ from Edit Page Effects? 2-5 What Kinds of Effects Does the Fusion Page Offer? 2-5 How Hard Is This Going to Be to Learn? 2-6 Chapter – 8 Introduction to Compositing in Fusion 2-3 DaVinci Resolve, Now With Fusion Inside The Fusion page is intended, eventually, to be a feature-complete integration of Blackmagic Design Fusion, a powerful 2D and 3D compositing application with over thirty years of evolution serving the film and broadcast industry, creating effects that have been seen in countless films and television series. Merged right into DaVinci Resolve with a newly updated user interface, the Fusion page makes it possible to jump immediately from editing right into compositing, with no need to export media, relink files, or launch another application to get your work done. Everything you need now lives right inside DaVinci Resolve. The Fusion page in DaVinci Resolve 15, showing Viewers, the Node Editor, and the Inspector How Do I Use the Fusion Page? At its simplest, you need only park the playhead over a clip you want to apply effects to, click the Fusion page button, and your clip is immediately available as a MediaIn node in the Fusion page, ready for you to add a variety of stylistic effects, paint out an unwanted blemish or feature, build a quick composite to add graphics or texture, or accomplish any other visual effect you can imagine, built from the Fusion page’s toolkit of effects. Alternately, you have the option of editing together all of the clips you want to use in your composition in the Edit page, superimposing and lining up every piece of media you’ll need with the correct timing, before selecting them and creating a Fusion clip, which functions as a single item in the Edit page timeline, but when seen in the Fusion page reveals each piece of media you’ve assembled as a fully built Fusion composition, ready for you to start adding tools to customize for whatever effect you need to create. Whichever way you want to work, all this happens on the very same timeline as editing, grading, and audio post, for a seamless back-and-forth as you edit, refine, and finish your projects. Chapter – 8 Introduction to Compositing in Fusion 2-4 How Do Fusion Effects Differ from Edit Page Effects? While there are many effects you can create in the Edit page, the Fusion page’s node-based interface has been designed to let you go deep into the minutiae of a composition to create sophisticated 2D and 3D effects with precise control and endless customization. If you like nodes for color correction, you’ll love them for effects. What Kinds of Effects Does the Fusion Page Offer? In addition to the kinds of robust compositing, paint, rotoscoping, and keying effects you’d expect from a fully-featured 2D compositing environment, the Fusion page offers much more. 3D Compositing The Fusion page has powerful 3D tools that include modeling text and simple geometry right inside the Fusion page. In Fusion Studio, this includes the ability to import 3D models in a variety of formats (that functionality has not yet been incorporated into DaVinci Resolve, but it’s coming). Once you’ve assembled a 3D scene, you can add cameras, lighting, and shaders, and then render the result with depth-of-field effects and auxiliary channels to integrate with more conventional layers of 2D compositing, for a sophisticated blending of 3D and 2D operations in the very same node tree. A 3D scene with textured 3D text, created entirely within the Fusion page Particles The Fusion page also has an extensive set of tools for creating particle systems that have been used in major motion pictures, with particle generators capable of spawning other generators, 3D particle generation, complex simulation behaviors that interact with 3D objects, and endless options for experimentation and customization, you can create particle system simulations for VFX, or more abstract particle effects for motion graphics. A 3D particle system, also created entirely within the Fusion page Chapter – 8 Introduction to Compositing in Fusion 2-5 Text The Text tools in the Fusion page are exceptional, giving you layout and animation options that DaVinci Resolve has never had before, in both 2D and 3D. Furthermore, these Text tools have been incorporated into the Edit page as Fusion Titles, which are compositions saved as macros with published controls, right in Fusion, that expose those controls in the Edit page Inspector for easy customization and control, even if you’re working with people who don’t know Fusion. A multi-layered text composite integrating video clips and Fusion-page generated elements And Lots More The list goes on, with Stereo and VR adjustment tools, Planar Tracking, Deep Pixel tools for re-compositing rendered 3D scenes using Auxiliary Channel data, powerful Masking and Rotoscoping tools, and Warping effects, the Fusion page is an impressively featured environment for building worlds, fixing problems, and flying multi-layered motion graphics animations through your programs. How Hard Is This Going to Be to Learn? That depends on what you want to do, but honestly, it’s not so bad with this PDF at your side, helping guide the way. It’s worth repeating that this Fusion documentation preview was developed specifically to help users who’ve never worked with Fusion before learn the core concepts needed to do the basics, in preparation for learning the rest of the application on your own. The Fusion page is another evolution of a deep, production-driven product that’s had decades of development, so its feature set is deep and comprehensive. You won’t learn it in an hour, but much of what you’ll find won’t be so very different from other compositing applications you may have used. And if you’ve familiarized yourself with the node-based grading workflow of the Color page, you’ve already got a leg up on understanding the central operational concept of compositing in the Fusion page. Go on, give it a try, and remember that you have the chapters of this PDF to refer to, which includes a lengthy “Learning to Work in the Fusion Page” section that walks you through a broad selection of basics, showing common techniques that you can experiment with using your own footage. Chapter – 8 Introduction to Compositing in Fusion 2-6 Chapter 9 Using the Fusion Page This chapter provides an orientation on the user interface of the Fusion page, providing a quick tour of what tools are available, where to find things, and how the different panels fit together to help you build and refine compositions in this powerful node-based environment. Chapter – 9 Using the Fusion Page 2-7 Contents The Fusion Page User Interface 2-10 The Work Area 2-11 Interface Toolbar 2-11 Viewers 2-12 Zooming and Panning into Viewers 2-13 Loading Nodes Into Viewers 2-14 Viewer Controls 2-14 Time Ruler and Transport Controls 2-16 The Playhead 2-16 The Current Time Field 2-16 Frame Ranges 2-17 Changing the Time Display Format 2-18 Zoom and Scroll Bar 2-18 Transport Controls 2-18 Keyframe Display in the Time Ruler 2-20 Fusion Viewer Quality and Proxy Options 2-20 The Fusion RAM Cache for Playback 2-21 Toolbar 2-22 Node Editor 2-23 Adding Nodes to Your Composition 2-24 Removing Nodes from Your Composition 2-24 Identifying Node Inputs and Node Outputs 2-25 Node Editing Essentials 2-25 Navigating the Node Editor 2-26 Keeping Organized 2-27 Tooltip Bar 2-27 Effects Library 2-28 Inspector 2-29 Keyframes Editor 2-31 Keyframe Editor Control Summary 2-31 Adjusting Clip Timings 2-32 Adjusting Effect Timings 2-32 Adjusting Keyframe Timings 2-32 Chapter – 9 Using the Fusion Page 2-8 Spline Editor 2-34 Spline Editor Control Summary 2-34 Choosing Which Parameters to Show 2-35 Essential Spline Editing 2-35 Essential Spline Editing Tools and Modes 2-36 Thumbnail Timeline 2-37 Media Pool 2-38 The Bin List 2-39 Importing Media Into the Media Pool on the Fusion Page 2-39 Bins, Power Bins, and Smart Bins 2-40 Showing Bins in Separate Windows 2-40 Filtering Bins Using Color Tags 2-41 Sorting the Bin List 2-42 Searching for Content in the Media Pool 2-42 The Console 2-43 Customizing the Fusion Page 2-44 The Fusion Settings Window 2-44 Showing and Hiding Panels 2-44 Resizing Panels 2-45 Chapter – 9 Using the Fusion Page 2-9 The Fusion Page User Interface If you open everything up at once, the Fusion page is divided into four principal regions designed to help you make fast work of node-based compositing. The Media Pool and Effects Library share the area found at the left, the Viewer(s) are at the top, the Work Area is at the bottom, and the Inspector at the right. All of these panels work together to let you add effects, paint to correct issues, create motion graphics or title sequences, or build sophisticated 3D and multi-layered composites, all without leaving DaVinci Resolve. The Fusion user interface shown completely However, the Fusion page doesn’t have to be that complicated, and in truth you can work very nicely with only the Viewer, Node Editor, and Inspector open for a simplified experience. A simplified set of Fusion controls for every-day working Chapter – 9 Using the Fusion Page 2-10 The Work Area You’ll probably not see this term used much, in favor of the specific panels within the work area that you’ll be using, but the area referred to as the Work Area refers to the region at the bottom half of the Fusion page UI, within which you can expose the three main panels used to construct compositions and edit animations in the Fusion page. These are the Node Editor, the Spline Editor, and the Keyframes Editor. By default, the Node Editor is the first thing you’ll see, and the main area you’ll be working within, but it can sit side-by-side with the Spline Editor and Keyframes Editor as necessary, and you can make more horizontal room on your display for these three panels by putting the Effects Library and Inspector into half-height mode, if necessary. The Work Area showing the Node Editor, the Spline Editor, and Keyframes Editor Interface Toolbar At the very top of the Fusion page is a toolbar with buttons that let you show and hide different parts of the Fusion page user interface. Buttons with labels identify which parts of the UI can be shown or hidden. If you right-click anywhere within this toolbar, you have the option of displaying this bar with or without text labels. The UI toolbar of the Fusion page These buttons are as follows, from left to right: Media Pool/Effects Library Full Height button: Lets you set the area used by the Media Pool and/or Effects Library to take up the full height of your display (you can display two of these UI areas at a time), giving you more area for browsing at the expense of a narrower Node Editor and Viewer area. At half height, the Media Pool/ Templates/Effects Library are restricted to the top half of the UI along with the Viewers (you can only show one at a time), and the Node Editor takes up the full width of your display. Media Pool: Shows and hides the Media Pool, from which you can drag additional clips into the Node Editor to use them in your Fusion page composition. Effects Library: Opens or hides the repository of all node tools that are available to use in the Fusion page. From here, you can click nodes to add them after the currently selected node in the Node Editor, or you can drag and drop nodes to any part of the node tree you like. Clips: Opens and closes the Thumbnail timeline, which lets you navigate your program, create and manage multiple versions of compositions, and reset the current composition. Chapter – 9 Using the Fusion Page 2-11 Nodes: Opens and closes the Node Editor, where you build and edit your compositions. Spline: Opens and closes the Spline Editor, where you can edit the curves that interpolate keyframe animation to customize and perfect their timing. Each keyframed parameter appears hierarchically within the effect in which it appears in a list to the left. Keyframes: Opens and closes the Keyframe Editor, which shows each clip and effects node in your Fusion composition as a layer. You can use the Keyframe Editor to edit and adjust the timing of keyframes that have been added to various effects in your composition. You can also use the Keyframe Editor to slide the relative timing of clips that have been added to the Fusion page, as well as to trim their In and Out points. A Spreadsheet can be shown and hidden within which you can numerically edit keyframe values for selected effects. Metadata: Hides or shows the Metadata Editor, which lets you read and edit the available clip and project metadata associated with any piece of media within a composite. Inspector: Shows or hides the Inspector, which shows you all the editable parameters and controls that correspond to selected nodes in the Node Editor. You can show the parameters for multiple nodes at once, and even pin the parameters of nodes you need to continue editing so that they’re displayed even if those nodes aren’t selected. Inspector Height button: Lets you open the Inspector to be half height (the height of the Viewer area) or full height (the height of your entire display). Half height lets you have more room for the Node Editor, Spline Editor, and/or Keyframes Editor, but full height lets you simultaneously edit more node parameters, or have enough room to display the parameters of multiple nodes at once. Viewers The Viewer Area can be set to display either one or two Viewers at the top of the Fusion page, and this is set via the Viewer button at the far right of the Viewer title bar. Each Viewer can show a single node’s output from anywhere in the node tree. You assign which node is displayed in which Viewer. This makes it easy to load separate nodes into each Viewer for comparison. For example, you can load a Keyer node into the left Viewer and the final composite into the right Viewer, so you can see the image you’re adjusting and the final result at the same time. Dual viewers let you edit an upstream node in one while seeing its effect on the overall composition in the other Chapter – 9 Using the Fusion Page 2-12 Ordinarily, each viewer shows 2D nodes from your composition as a single image. However, when you’re viewing a 3D node, you have the option to set that viewer to one of several 3D views, including a perspective view that gives you a repositionable stage on which to arrange the elements of the world you’re creating, or a quad-view that lets you see your composition from four angles, making it easier to arrange and edit objects and layers within the XYZ axes of the 3D space in which you’re working. Loading a 3D node into a Viewer switches on a perspective view TIP: In perspective view, you can hold the Option key down and drag in the viewer to pivot the view around the center of the world. All other methods of navigating Viewers work the same. Similarly to the Color page, the Viewers have a variety of capabilities you can use to compare and evaluate what you’re looking at, except that there are many more options that are specific to the detail-oriented work compositing entails. This section gives a short overview of viewer capabilities to get you started. Zooming and Panning into Viewers There are standardized methods of zooming into and panning around Viewers when you need a closer look at the situation. These methods also work with the Node Editor, Spline Editor, and Keyframes Editor. Methods of navigating Viewers: Middle click and drag to pan around the Viewer. Hold Shift and Command down and drag the Viewer to pan. Press the Middle and Left buttons simultaneously and drag to resize the Viewer. Hold the Command key down and use your pointer’s scroll control to resize the Viewer. In 3D perspective view, hold the Option key down and drag to spin the stage around. Chapter – 9 Using the Fusion Page 2-13 Loading Nodes Into Viewers When you first open the Fusion page, the output of the current empty composition (the MediaOut1 node) is usually showing in Viewer2. If you’re in Dual-viewer mode, Viewer1 remains empty until you assign a node to one of them. To load specific nodes into specific viewers: Hover the pointer over a node, and click one of two buttons that appear at the bottom-left of the node. Click once to select a node, and press 1 (for the left Viewer) or 2 (for the right Viewer). Right-click a node and choose View On > None/LeftView/RightView in the contextual menu. Drag a node and drop it over the Viewer you’d like to load it into (this is great for tablet users). When a node is being viewed, a Viewer Assignment button appears at the bottom-left. This is the same control that appears when you hover the pointer over a node. Not only does this control let you know which nodes are loaded into which viewer, but they also expose little round buttons for changing which viewer they appear in. Viewer assignment buttons at the bottom left of nodes indicate when they’re being viewed, and which dot is highlighted indicates which Viewer that node is loaded into Viewer Controls A series of buttons and pop-up menus in the Viewer’s title bar provides several quick ways of customizing the Viewer display. Controls in the Viewer title bar Zoom menu: Lets you zoom in on the image in the Viewer to get a closer look, or zoom out to get more room around the edges of the frame for rotoscoping or positioning different layers. Choose Fit to automatically fit the overall image to the available dimensions of the Viewer. Split Wipe button and A/B Buffer menu: You can actually load two nodes into a single viewer using that viewer’s A/B buffers by choosing a buffer from the menu and dragging a node into the Viewer. Turning on the Split Wipe button (press Forward Slash) shows a split wipe between the two buffers, which can be dragged left or right via the handle of the on-screen control, or rotated by dragging anywhere on the dividing line on the on-screen control. Alternately, you can switch between each full-screen buffer to compare them (or to dismiss a split-screen) by pressing Comma (A buffer) and Period (B buffer). Chapter – 9 Using the Fusion Page 2-14 SubView type: (these aren’t available in 3D viewers) Clicking the icon itself enables or disables the current “SubView” option you’ve selected, while using the menu lets you choose which SubView is enabled. This menu serves one of two purposes. When displaying ordinary 2D nodes, it lets you open up SubViews, which are viewer “accessories” within a little pane that can be used to evaluate images in different ways. These include an image Navigator (for navigating when zoomed way into an image), Magnifier, 2D Viewer (a mini-view of the image), 3D Histogram scope, Color Inspector, Histogram scope, Image Info tooltip, Metadata tooltip, Vectorscope, or Waveform scope. The Swap option (Shift-V) lets you switch what’s displayed in the Viewer with what’s being displayed in the Accessory pane. When displaying 3D nodes, this button lets you turn on the quad-paned 3D Viewer. Node name: The name of the currently viewed node is displayed at the center of the Viewer’s title bar. RoI controls: Clicking the icon itself enables or disables RoI limiting in the Viewer, while using the menu lets you choose the region of the RoI. The Region of Interest (RoI) lets you define the region of the Viewer in which which pixels actually need to be rendered. When a node renders, it intersects the current RoI with the current Domain of Definition (DoD) to determine what pixels should be affected. When enabled, you can position a rectangle to restrict rendering to a small region of the image, which can significantly speed up performance when you’re working on very high resolution or complex compositions. Auto (the default) sets the region to whatever is visible at the current zoom/pan level in the Viewer. Choosing Set lets you draw a custom region within the frame by dragging a rectangle that defaults to the size of the Viewer, which is resizable by dragging the corners or sides of the on-screen control. Choosing Lock prevents changes from being made to the current RoI. Choosing Reset resets the RoI to the whole Viewer. Color controls: Lets you choose which color and/or image channels to display in the Viewer. Clicking the icon itself toggles between Color (RGB) and Alpha, the two most common things you want to see (pressing C also toggles between Color and Alpha). Opening the menu displays every possible channel that can be displayed for the currently viewed node, commonly including RGB, Red, Green, Blue, Alpha (available from the keyboard by pressing R, G, or B). For certain media and nodes, additional auxiliary channels are available to be viewed, including Z-depth, Object ID, Material ID, XYZ Normals, etc. Viewer LUT: Clicking the icon itself toggles LUT display on or off, while the menu lets you choose which of the many available color space conversions to apply. By default, Viewers in Fusion show you the image prior to any grading done in the Color page, since the Fusion page comes before the Color page in the DaVinci Resolve image processing pipeline. However, if you’re working on clips that have been converted to linear color space for compositing, you may find it desirable to composite and make adjustments to the image relative to a normalized version of the image that appears close to what the final will be, and enabling the LUT display lets you do this as a preview, without permanently applying this color adjustment to the image. The top five options let you choose Fusion controls, which can be customized via the Edit item at the bottom of this menu. The rest of this menu shows all LUTs from the /Library/ Application Support/Blackmagic Design/DaVinci Resolve/LUT/VFX IO/ directory (on macOS) to use for viewing. Chapter – 9 Using the Fusion Page 2-15 Option menu: This menu contains various settings that pertain to the Viewer in the Fusion page. –– Checker Underlay: Toggles a checkerboard underlay that makes it easy to see areas of transparency. –– Show Controls: Toggles whatever onscreen controls are visible for the currently selected node. –– Pixel Grid: Toggles a preview grid that shows, when zoomed in, the actual size of pixels in the image. Time Ruler and Transport Controls The Time Ruler, located beneath the Viewer Area, shows the frame range of the current clip or composition. However, the duration of this range depends on what’s currently selected in the Timeline: If you’ve selected a clip, then the Time Ruler displays all source frames for that clip, and the current In and Out points for that clip define the “render range,” or the range used in the Timeline and thus available in the composition by default. All frames outside of this range constitute the heads and tails of that clip that are unused in the edited Timeline. The Time Ruler displaying ranges for a clip in the Timeline via yellow marks (the playhead is red) If you’ve selected a Fusion clip or a compound clip, then the “working range” reflects the entire duration of that clip. The Time Ruler displaying ranges for a Fusion clip in the Timeline The Playhead A red playhead within the Time Ruler indicates the currently viewed frame. Clicking anywhere within the Time Ruler jumps the playhead to that frame, and dragging within the Time Ruler drags the playhead within the available duration of that clip or composition. The Current Time Field The Current Time field at the right of the Transport controls shows the frame at the position of the playhead, which corresponds to the frame seen in the Viewer. However, you can also enter time values into this field to move the playhead by specific amounts. When setting ranges and entering frame numbers to move to a specific frame, numbers can be entered in subframe increments. You can set a range to be -145.6 to 451.75 or set the Playhead to 115.22. This can be very helpful to set keyframes where they actually need to occur, rather Chapter – 9 Using the Fusion Page 2-16 than on a frame boundary, so you get more natural animation. Having subframe time lets you use time remapping nodes or just scale keyframes in the Spline view and maintain precision. NOTE: Many fields in the Fusion page can evaluate mathematical expressions that you type into them. For example, typing 2 + 4 into most fields results in the value 6.0 being entered. Because Feet + Frames uses the + symbol as a separator symbol, the Current Time field will not correctly evaluate mathematical expressions that use the + symbol, even when the display format is set to Frames mode. Frame Ranges The Time Ruler uses two different frame ranges, one for the duration of the entire clip or composition, and a Render range that currently determines either the duration of the current clip that appears within the Timeline, or the range of frames to cache in memory for previews. Composition Start and End Range The Composition Start and End range is simply the total duration of the current composition. Render Range The Render Start and End range determines the range of frames that will be used for interactive playback, disk caches, and previews. The range is normally visible in the time slider as a light gray highlighted region within the Time Ruler. Frames outside the Render range will not be rendered or played, although you can still drag the Playhead or set the current time to these frames to see what the image looks like. Two fields at the far left of the Transport controls show the first frame and last frame of this range. You can modify the Render range in a variety of ways. You can set the Render range in the Time Ruler by doing one of the following: Hold the Command key down and drag a new range within the Time Ruler. Right-click within the Time Ruler and choose Set Render Range from the contextual menu. Enter new ranges in the Range In and Out fields to the left of the Transport controls. Drag a node from the Node Editor to the Time Ruler to set the range to the duration of that node. The Render Start and Render End time fields Chapter – 9 Using the Fusion Page 2-17 Changing the Time Display Format By default, all time fields and markers in the Fusion Page count in frames, but you can also set time display to SMPTE timecode or Feet + Frames. To change the time display format: 1 Choose Fusion > Fusion Settings. 2 When the Fusion settings dialog opens select the Defaults panel and choose a Timecode option. 3 Open the Frame Format panel. If you’re using timecode, choose a Frame Rate and turn on the “has fields” checkbox if your project is interlaced. If you’re using feet and frames, set the Film Size value to match the number of frames found in a foot of film in the format used in your project. 4 Click Save. Zoom and Scroll Bar A two-handled scroll bar lets you zoom into the range shown by the Time Ruler, which is useful if you’re looking at a clip with really long handles such that the Render range is a tiny sliver in the Time Ruler. Dragging the left or right handles of this bar zooms relative to the opposite handle, enlarging the width of each displayed frame. Once you’ve zoomed in, you can drag the scroll bar left or right to scroll through the composition. TIP: Holding the middle mouse button and dragging in the Time Ruler lets you scroll the visible range. Transport Controls There are six Transport controls underneath the Time Ruler, including Composition first frame, Play Reverse, Stop, Play Forward, Composition last frame, and Loop. The Fusion page Transport controls Navigation Shortcuts Many of the standard Transport control keyboard shortcuts also work in the Fusion page, but some are specific to Fusion’s particular needs. To move the playhead in the Time Ruler using the keyboard, do one of the following: Space Bar: Toggles forward playback on and off. JKL: Basic JKL playback is supported, including J to play backward, K to stop, and L to play forward. Back Arrow: Moves 1 frame backward. Forward Arrow: Moves 1 frame forward. Shift-Back Arrow: Moves to the clip’s Source End Frame. Chapter – 9 Using the Fusion Page 2-18 Shift-Forward Arrow: Moves to the clip’s Source Start Frame. Command-Back Arrow: Jumps to the clip’s In Point. Command-Forward Arrow: Jumps to the clip’s Out Point. Real Time Playback Not Guaranteed Because many of the effects you can create in the Fusion page are processor-intensive, there is no guarantee of real time playback at your project’s full frame rate, unless you’ve cached your composition first (see later). Frame Increment Options Right-clicking either the Play Reverse or Play Forward buttons opens a contextual menu with options to set a frame increment value, which lets you move the playhead in sub-frame or multi-frame increments whenever you use a keyboard shortcut to move frame by frame through a composition. Moving the playhead in multi-frame increments can be useful when rotoscoping. Moving the playhead in sub-frame increments can be useful when rotoscoping or inspecting interlaced frames one field at a time (0.5 of a frame). Right-click the Play Forward or Play Backwards buttons to choose a frame increment in which to move the playhead Looping Options The Loop button can be toggled to enable or disable looping during playback. You can rightclick this button to choose the looping method that’s used: Playback Loop: The playhead plays to the end of the Time Ruler and starts from the beginning again. Pingpong Loop: When the playhead reaches the end of the Time Ruler, playback reverses until the playhead reaches the beginning of the Time Ruler, and then continues to ping pong back and forth. Chapter – 9 Using the Fusion Page 2-19 Keyframe Display in the Time Ruler When you select a node that’s been animated with keyframed parameters, those keyframes appear in the Time Ruler as little white tic marks, letting you navigate among and edit keyframes without being required to open the Keyframes Editor or Spline Editor to see them. The Time Ruler displaying keyframe marks To move the playhead in the Time Ruler among keyframes: Press Option-Left Bracket ([) to jump to the next keyframe to the left. Press Option-Right Bracket (]) to jump to the next keyframe to the right. Fusion Viewer Quality and Proxy Options Right-clicking anywhere in the Transport control area lets you turn on and off Fusion pagespecific quality controls, which lets you either enable high-quality playback at the expense of greater processing times, or enter various proxy modes that temporarily lower the display quality of your composition in order to speed processing as you work. Rendering for final output is always done at the highest quality, regardless of these settings. High Quality As you build a composition, often the quality of the displayed image is less important than the speed at which you can work. The High Quality setting gives you the option to either display images with faster interactivity or at final render quality. When you turn off High Quality, complex and time consuming operations such as area sampling, anti-aliasing, and interpolation are skipped to render the image to the Viewer more quickly. Enabling High Quality forces a full quality render to the Viewer that’s identical to what will be output during final delivery. Motion Blur Turning Motion Blur off temporarily disables motion blur throughout the composition, regardless of any individual nodes for which it’s enabled. This can significantly speed up renders to the Viewer. Proxy A draft mode to speed processing while you’re building your composite. Turning on Proxy reduces the resolution of the images that are rendered to the Viewer, speeding render times by causing only one out of every x pixels to be processed, rather than processing every pixel. The value of x is decided by adjusting a slider in the General panel of the Fusion Settings, found in the Fusion menu. Auto Proxy A draft mode to speed processing while you’re building your composite. Turning on Auto Proxy reduces the resolution of the image while you click and drag on a parameter’s control to make an adjustment. Once you release that control, the image snaps back to its original resolution. This lets you adjust processor intensive operations more smoothly, without the wait for every frame to render at full quality causing jerkiness. You can set the auto proxy ratio by adjusting a slider in the General panel of the Fusion Settings, found in the Fusion menu. Chapter – 9 Using the Fusion Page 2-20 Selective Updates (Available in Fusion Settings) There are three options: Update All: Forces all of the nodes in the current node tree to render. This is primarily used when you want to update all of the thumbnails displayed in the Node Editor. Selective: (the default) Causes only nodes that directly contribute to the current image to be rendered, so named because only selective nodes are rendered. No Update: Prevents rendering altogether, which can be handy for making a lot of changes to a slow-to-render composition. While set to None, the Node Editor, Keyframes Editor and Spline Editor will be highlighted with a red border to indicate that the Tools are not being updated. The Fusion RAM Cache for Playback When assembling a node tree, all image processing operations are rendered live to display the final result in the Viewers. However, as each frame is rendered, and especially as you initiate playback forward or backward, these images are automatically stored to a RAM cache as they’re processed so you can replay those frames in real time. The actual frame rate achieved during playback is displayed in the Tooltip bar at the bottom of the Fusion page during playback. Of course, when you play beyond the cached area of the Time Ruler, uncached frames will need to be rendered before being added to the cache. Priority is given to caching nodes that are currently being displayed, based on which nodes are loaded to which Viewers. However, other nodes may also be cached, depending on available memory and on how processor-intensive those nodes happen to be, among other factors. Memory Limits of the RAM Cache When the size of the cache reaches the Fusion Memory Limits setting found in the Configuration panel of the System Preferences, then lower-priority cache frames are automatically discarded to make room for new caching. You can keep track of how much of the RAM cache has been used via a percentage of use indicator at the far right of the Tooltip bar at the bottom of the Fusion page. Percentage of the RAM cache that’s been used at the bottom right of the Fusion page Displaying Cached Frames All frames that are cached for the currently viewed range of nodes are indicated by a green line at the bottom of the Time Ruler. Any green section of the Time Ruler should pay back in real time. The green line indicates frames that have been cached for playback Chapter – 9 Using the Fusion Page 2-21 Temporarily Preserving the Cache When Changing Quality or Proxy Settings If you toggle the composition’s quality settings or proxy options, the cache is not immediately discarded; the green line instead turns red to let you know the cache is being preserved and can be used again when you go back to the original level of quality or disable proxy mode. However, if you play through those frames at the new quality or proxy settings, this preserved cache will be overwritten with a new cache at the current quality or proxy setting. A red line indicates that cached frames from a different quality or proxy setting are being preserved There’s one exception to this, however. When you cache frames at the High Quality setting and you then turn High Quality off, the green frames won’t turn red. Instead, the High Quality cached frames will be used even though the HiQ setting has been disabled. Toolbar The Toolbar, located underneath the Time Ruler, contains buttons that let you quickly add commonly used nodes to the Node Editor. Clicking any of these buttons adds that node after the currently selected node in the node tree, or adds an unconnected instance of that node if no nodes are selected. The Toolbar has buttons for adding commonly used nodes to the Node Editor The Toolbar is divided into six sections that group commonly used nodes together. As you hover the pointer over any button, a tooltip shows you that node’s name. Generator/Title/Paint nodes: The Background and FastNoise generators are commonly used to create all kinds of effects, and the Title generator is obviously a ubiquitous tool, as is Paint. Color/Blur nodes: ColorCorrector, ColorCurves, HueCurves, and BrightnessContrast are the four most commonly used color adjustment nodes, while the Blur node is ubiquitous. Compositing/Transform nodes: The Merge node is the primary node used to composite one image against another. ChannelBooleans and MatteControl are both essential for re-assigning channels from one node to another. Resize alters the resolution of the image, permanently altering the available resolution, while Transform applies pan/tilt/rotate/zoom effects in a resolution-independent fashion that traces back to the original resolution available to the source image. Mask nodes: Rectangle, Ellipse, Polygon, and BSpline mask nodes let you create shapes to use for rotoscoping, creating garbage masks, or other uses. Particle system nodes: Three particle nodes let you create complete particle systems when you click them from left to right. PEmitter emits particles in 3D space, while pMerge lets you merge multiple emitters and particle effects to create more complex systems. pRender renders a 2D result that can be composited against other 2D images. Chapter – 9 Using the Fusion Page 2-22 3D nodes: Seven 3D nodes let you build sophisticated 3D scenes. These nodes auto attach to one another to create a quick 3D template when you click from left to right. ImagePlane3D lets you connect 2D stills and movies for compositing into 3D scenes. Shape3D lets you create geometric primitives of different kinds. Text3D lets you build 3D text objects. Merge3D lets you composite multiple 3D image planes, primitive shapes, and 3D text together to create complex scenes, while SpotLight lets you light the scenes in different ways, and Camera3D lets you frame the scene in whatever ways you like. Renderer3D renders the final scene and outputs 2D images and auxiliary channels that can be used to composite 3D output against other 2D layers. When you’re first learning to use Fusion, these nodes are really all you need to build most common composites, but even once you’ve become a more advanced user, you’ll still find that these are truly the most common operations you’ll use. Node Editor The Node Editor is the heart of the Fusion page, because it’s where you build the tree of nodes that makes up each composition. Each node you add to the node tree adds a specific operation that creates one effect, whether it’s blurring the image, adjusting color, painting strokes, drawing and adding a mask, extracting a key, creating text, or compositing two images into one. You can think of each node as a layer in an effects stack, except that you have the freedom to route image data in any direction to branch and merge different segments of your composite in completely nonlinear ways. This makes it easy to build complex effects, but it also makes it easy to see what’s happening, since the node tree doubles as a flowchart that clearly shows you everything that’s happening, once you learn to read it. The Node Editor displaying a node tree creating a composition Chapter – 9 Using the Fusion Page 2-23 Adding Nodes to Your Composition Depending on your mood, there are a few ways you can add nodes from the Effects Library to your composition. For most of these methods, if there’s a single selected node in the Node Editor, new nodes are automatically added to the node tree after it, but if there are no selected nodes or multiple selected nodes, then new nodes are added as disconnected from anything else. Methods of adding nodes: Click a button in the tool bar. Open the Effects Library, find the node you want in the relevant category, and click once on a node you’d like to add. Right-click on a node and choose Insert Tool from the contextual menu to add it after the node you’ve right-clicked on. Or, you can right-click on the background of the Node Editor to use that submenu to add a disconnected node. Press Shift-Spacebar to open a Select Tool dialog, type characters corresponding to the name of the node you’re looking for, and press the Return key (or click OK) when it’s found. Once you learn this method, it’ll probably become one of your most frequentlyused ways of adding nodes. The Select Tool dialog lets you find any node quickly if you know its name Removing Nodes from Your Composition Removing nodes is as simple as selecting one or more nodes, and then pressing the Delete or Backspace keys. Chapter – 9 Using the Fusion Page 2-24 Identifying Node Inputs and Node Outputs If you hover the pointer over any of a node’s inputs or outputs, the name of that input or output will immediately appear in the Tooltip bar, and if you wait for a few more moments, a floating tooltip will display the same name right over the node you’re working on. Node Editing Essentials Each node has inputs and outputs that are “wired together” using connections. The inputs are represented by arrows that indicate the flow of image data from one node to the next, as each node applies its effect and feeds the result (via the square output) to the next node in the tree. In this way, you can quickly build complex results from a series of relatively simple operations. Three nodes connected together You can connect a single node’s output to the inputs of multiple nodes (called “branching”). One node branching to two to split the image to two operations You can then composite images together by connecting the output from multiple nodes to certain nodes such as the Merge node that combine multiple inputs into a single output. Two nodes being merged together into one to create a composite Chapter – 9 Using the Fusion Page 2-25 By default, new nodes are added from left to right in the Node Editor, but they can also flow from top to bottom, left to right, bottom to top, or in all directions simultaneously. Connections automatically reorient themselves along all four sides of each node to maintain the cleanest possible presentation as you rearrange other connected nodes. Nodes can be oriented in any direction; the input arrows let you follow the flow of image data Navigating the Node Editor As your composition gets larger, parts of it will inevitably go off screen. By default, when a portion of the node tree has gone off-screen, a resizable Navigator pane appears at the upper right corner, which can be used to see a miniature representation of the entire node tree that you can drag within to pan to different parts of your composition quickly. You can resize the navigator using a handle at the lower left-hand corner, and you can choose to show or hide the navigator by right-clicking the Node Editor to access the Options submenu of the contextual menu. The Navigator pane for accessing offscreen parameters or tools There are other standard methods of panning and zooming around the Node Editor. Chapter – 9 Using the Fusion Page 2-26 Methods of navigating the Node Editor: Middle click and drag to pan around the Node Editor. Hold Shift and Command down and drag the Node Editor to pan. Press the Middle and Left buttons simultaneously and drag to resize the Node Editor. Hold the Command key down and use your pointer’s scroll control to resize the Node Editor. Right-click the Node Editor and choose an option from the Scale submenu of the contextual menu. Press Command-1 to reset the Node Editor to its default size. Keeping Organized As you work, it’s important to keep the node trees that you create tidy to facilitate a clear understanding of what’s happening. Fortunately, the Fusion page Node Editor provides a variety of methods and options to help you with this, found within the Options and Arrange Tools submenus of the Node Editor contextual menu. Tooltip Bar The Tooltip bar at the bottom of the Fusion page, immediately above the Resolve Page bar, shows you a variety of up-to-date information about things you’re selecting and what’s happening in the Fusion page. For example, hovering the pointer over any node displays information about that node in the Tooltip bar (as well as in a floating tooltip), while the currently achieved frame rate appears whenever you initiate playback, and the percentage of the RAM cache that’s used appears at all times. Other information, updates, and warnings appears in this area as you work. The tooltip bar under the Node Editor showing you information about a node under the pointer Occasionally the Tooltip bar will display a badge to let you know there’s a message in the console you might be interested in. The badge will indicate if the message is an error, log, or script message. A notification that there’s a message in the Console Chapter – 9 Using the Fusion Page 2-27 Effects Library The Effects Library on the Fairlight page is currently restricted to displaying only the nodes and effects that are available in the Fusion page. ResolveFX and third-party OFX are not able to be used in the Fusion page at this time, although that capability will be added eventually. While the Toolbar shows many of the most common nodes you’ll be using in any composite, the Effects Library contains every single tool available in the Fusion page, organized by category, with each node ready to be quickly added to the Node Editor. Suffice it to say there are many, many more nodes available in the Effects Library than on the Toolbar, spanning a wide range of uses. The Effects Library with Tools open The hierarchical category browser of the Effects Library is divided into two sections. The Tools section contains every node that represents an elemental image processing operation in the Fusion page. The Templates section contains a variety of additional compositing functions, as well as libraries of content such as Lens Flares, Backgrounds, Generators, Particle Systems, Shaders (for texturing 3D objects) and other resources for use in your composites. Chapter – 9 Using the Fusion Page 2-28 The Templates section of the Effects Library Similar to the Media Pool, the Effects Library’s bin list can be made full-height or half-height using a button at the far left of the UI toolbar. Additionally, an Option menu in the Effects Library gives you access to additional options and commands. Inspector The Inspector is a panel on the right side of the Fusion page that you use to display and manipulate the parameters of one or more selected nodes. When a node is selected in the Node Editor, its parameters and settings appear in the Inspector. The Inspector shows parameters from one or more selected nodes The Tools and Modifiers Panels The Fusion Inspector is divided into two panels. The Tools panel shows you the parameters of selected nodes. The Modifiers panel shows you different things for different nodes. For all nodes, it shows you the controls for Modifiers, or adjustable expressions, that you’ve added to Chapter – 9 Using the Fusion Page 2-29 specific parameters to automatically animate them in different ways. In the following image, a Perturb modifier has been added to a parameter to add random animation to that parameter, and the controls found on the Modifier panel lets you customize what kind of randomness is being added. The Modifier panel showing a Perturb modifier Other nodes display more specific items here. For example, Paint nodes show each brush stroke as an individual set of controls in the Modifiers panel, available for further editing or animating. Parameter Header Controls A cluster of controls appears at the top of every node’s controls in the Inspector. Common Inspector Controls Set Color: A pop-up menu that lets you assign one of 16 colors to a node, over-riding a node’s own color. Versions: Clicking Versions reveals another toolbar with six buttons. Each button can hold an individual set of adjustments for that node that you can use to store multiple versions of an effect. Pin: The Fusion page Inspector is also capable of simultaneously displaying all parameters for multiple nodes you’ve selected in the Node Editor. Furthermore, a Pin button in the title bar of each node’s parameters lets you “pin” that node’s parameters into the Inspector so that they remain there even when that node is de-selected, which is valuable for key nodes that you need to adjust even while inspecting other nodes of your composition. Lock: Locks that node so that no changes can be made to it. Reset: Resets all parameters within that node. Parameter Tabs Many nodes expose multiple tabs worth of controls in the Inspector, seen as icons at the top of the parameter section for each node. Click any tab to expose that set of controls. Nodes with several tabs worth of parameters Chapter – 9 Using the Fusion Page 2-30 Keyframes Editor The Keyframes Editor displays each MediaIn and effects node in the current composition as a stack of layers within a miniature timeline. The order of the layers is largely irrelevant as the order and flow of connections in the node tree dictates the order of image processing operations. You can use the Keyframes Editor to trim, extend, or slide MediaIn and effects nodes, or to adjust the timing of keyframes, which appear superimposed over each effect node unless you open them up into their own editable track. The Keyframes Editor is used to adjust the timing of clips, effects, and keyframes Keyframe Editor Control Summary At the top, a series of zoom and framing controls let you adjust the work area containing the layers. Vertical and horizontal zoom controls let you scale the size of the editor. A Zoom to Fit button fits the width of all layers to the current width of the Keyframes Editor. A Zoom to Rect tool lets you draw a rectangle to define an area of the Keyframe Editor to zoom into. A Sort pop-up menu lets you sort or filter the layers in various ways. An Option menu provides access to many other ways of filtering layers and controlling visible options. A timeline ruler provides a time reference, as well as a place in which you can scrub the playhead. At the left, a track header contains the name of each layer, as well as controls governing that layer. A lock button lets you prevent a particular layer from being changed. Nodes that have been keyframed have a disclosure control, which when opened displays a keyframe track for each animated parameter. In the middle, the actual editing area displays all layers and keyframe tracks available in the current composition. Chapter – 9 Using the Fusion Page 2-31 At the bottom-left, Time Stretch and Spreadsheet mode controls provide additional ways to manipulate keyframes. At the bottom-right, the Time/Toffset/Tscale pop-up menu and value fields let you numerically alter the position of selected keyframed either absolutely, relatively, or to a scale. Adjusting Clip Timings Each MediaIn node that represents a clip used in a composition is represented as a layer in this miniature timeline. You can edit a layer’s In or Out points by positioning the pointer over the beginning or end of a clip and using the resize cursor to drag that point to a new location. You can slide a layer by dragging it to the left or right, to better line up with the timing of other layers in your composition. While much of this could be done in the Timeline prior to creating a Fusion clip that contains several MediaIn nodes, the Keyframes Editor also lets you adjust the timing of clips that you’ve added from directly within the Fusion page, as well as generators and 3D objects, which never originally appeared in the Edit page Timeline at all. Adjusting Effect Timings Each Effect node also appears as a layer, just like clips. You can resize the In and Out points of an Effect layer, and slide the entire layer forward or backward in time, just like MediaIn layers. If you trim an Effects layer to be shorter than the duration of the composition, the effect will cut in at whichever frame the layer begins, and cut out at after the last frame of that layer, just like a clip on a timeline. Adjusting Keyframe Timings When you’ve animated an effect by adding keyframes to a parameter in the Inspector, the Keyframes Editor is used to edit the timing of keyframes in a simple way. By default, all keyframes applied to parameters within a particular node’s layer appear superimposed in one flat track over the top of that layer. To edit keyframes, you can click the disclosure control to the left of any animated layer’s name in the track header, which opens up keyframe tracks for every keyframed parameter within that layer. Keyframe tracks exposed Keyframe Editing Essentials Here’s a short list of keyframe editing methods that will get you started. Methods of adjusting keyframes: You can click on a single keyframe to select it. You can drag a bounding box over a series of keyframes to select them all. You can drag keyframes left and right to reposition them in time. You can right-click one or more selected keyframes and use contextual menu commands to change keyframe interpolation, copy/paste keyframes, or even create new keyframes. Chapter – 9 Using the Fusion Page 2-32 To change the position of a keyframe using the Toolbar, do one of the following: Select a keyframe, then enter a new frame number in the Time Edit box. Select a keyframe(s), click the Time button to switch to Time Offset mode, then enter a frame offset. Select a keyframe(s), click the Time button twice to switch to T Scale mode, then enter a frame offset. Time Stretching Keyframes If you select a range of keyframes in a keyframe track, you can turn on the Time Stretch tool to show a box you can use to squeeze and stretch the entire range of keyframes relative to one another, to change the overall timing of a sequence of keyframes without losing the relative timing from one keyframe to the next. Alternately, you can turn on Time Stretch and draw a bounding box around the keyframes you want to adjust to create a time stretching boundary that way. Click the Time Stretch tool again to turn it off, Time Stretching keyframes The Keyframe Spreadsheet If you turn on the Spreadsheet and then click on the name of a layer in the a keyframe track, the numeric time position and value (or values if it’s a multi-dimensional parameter) of each keyframe appear as entries in the cells of the Spreadsheet. Each column represents one keyframe, while each row represents a single aspect of each keyframe. Editing keyframes in the Spreadsheet For example, if you’re animating a blur, then the “Key Frame” row shows the frame each keyframe is positioned at, and the “Blur1BlurSize” row shows the blur size at each keyframe. If you change the Key Frame value of any keyframe, you’ll move that keyframe to a new frame of the Timeline. Chapter – 9 Using the Fusion Page 2-33 Spline Editor The Spline Editor provides a more detailed environment for editing the timing and value of keyframes that create different animated effects, using control points at each keyframe connected by splines (also called curves) that let you adjust how animated values change over time. The Spline Editor has four main areas, the Zoom and Framing controls at top, the Parameter list at the left, the Graph Editor in the middle, and the Toolbar at the bottom. The Spline Editor is divided into the Zoom controls at top, Parameter list at left, and Toolbar Spline Editor Control Summary At the top, a series of Zoom and Framing controls let you adjust the work area containing the layers. Vertical and horizontal zoom controls let you scale the size of the editor. A Zoom to Fit button fits the width of all layers to the current width of the Keyframes Editor. A Zoom to Rect tool lets you draw a rectangle to define an area of the Keyframe Editor to zoom into. A Sort pop-up menu lets you sort or filter the layers in various ways. An Option menu provides access to many other ways of filtering layers and controlling visible options. A timeline ruler provides a time reference, as well as a place in which you can scrub the playhead. The Parameter list at the left is where you decide which splines are visible in the Graph view. By default, the Parameter list shows every parameter of every node in a hierarchical list. Checkboxes beside each name are used to show or hide the curves for different keyframed parameters. Color controls let you customize each spline’s tint, to make splines easier to see in a crowded situation. Chapter – 9 Using the Fusion Page 2-34 The Graph view that takes up most of this panel shows the animation spline along two axes. By default, the horizontal axis represents time and the vertical axis represents the spline’s value, although you can change this via the Horizontal and Vertical Axis pop-up menus at the bottom-right of the Spline Editor, and selected control points show their values in the accompanying edit fields. Lastly, the toolbar at the bottom of the Spline Editor has controls to set control point interpolation, spline looping, or choose Spline editing tools for different purposes. Choosing Which Parameters to Show Before you start editing splines to customize or create animation, you need to choose which parameter’s splines you want to work on. To show every parameter in every node: Click the Splines Editor Option menu and choose Expose All Controls. Toggle this control off again to go back to viewing what you were looking at before. To show splines for the currently selected node: Click the Splines Editor Option menu and choose Show Only Selected Tool. Essential Spline Editing The Spline Editor is a deep and sophisticated environment for keyframe and spline editing and retiming, but the following overview will get you started using this tool for creating and refining animation. To select one or more control points: Click any control point to select it. Command-click multiple control points to select them. Drag a bounding box around multiple control points to select them as a group. To edit control points and splines: Click anywhere on a spline to add a control point. Drag one or more selected control points to reshape the spline. Shift-drag a control point to constrain its motion vertically or horizontally. To edit Bezier curves: Select any control point to make its Bezier handles visible, and drag the Bezier handles. Command-drag a Bezier handle to break the angle between the left and right handles. To delete control points: Select one or more control points and press the delete or backspace key. Chapter – 9 Using the Fusion Page 2-35 Essential Spline Editing Tools and Modes The Spline Editor toolbar at the bottom contains a mix of control point interpolation buttons, Spline loop modes, and Spline editing tools. Control Point Interpolation The first five buttons let you adjust the interpolation of one or more selected control points. Control point interpolation controls Smooth: Creates automatically adjusted Bezier curves to create smoothly interpolating animation. Flat: Creates linear control points. Invert: Inverts the vertical position of selected keyframes relative to one another. Step In: For each keyframe, creates sudden changes in value at the next keyframe to the right. Similar to a hold keyframe in After Effects, or a static keyframe in the Color page. Step Out: Creates sudden changes in value at every keyframe for which there’s a change in value at the next keyframe to the right. Similar to a hold keyframe in After Effects, or a static keyframe in the Color page. Reverse: Reverses the horizontal position of selected keyframes in time, so the keyframes are backwards. Spline Loop Modes The next three buttons let you set up spline looping after the last control point on a parameter’s spline, enabling a limited pattern of keyframes to animate over a far longer duration. Only the control points you’ve selected are looped. Spline Loop modes Set Loop: Repeats the same pattern of keyframes over and over. Set Ping Pong: Repeats a reversed set of the selected keyframes and then a duplicate set of the selected keyframes to create a more seamless pattern of animation. Set Relative: Repeats the same pattern of selected keyframes but with the values of each repeated pattern of keyframes being incremented or decremented by the trend of all keyframes in the selection. This results in a loop of keyframes where the value either steadily increases or decreases with each subsequent loop. Spline Editing Tools The next five buttons provide specialized Spline editing tools. Spline editing controls Chapter – 9 Using the Fusion Page 2-36 Select All: Selects every keyframe currently available in the Splines Editor. Click Append: Click once to select this tool, click again to de-select it. Lets you add or adjust keyframes and spline segments (sections of splines between two keyframes) depending on the keyframe mode you’re in. With Smooth or Linear keyframes, clicking anywhere above or below a spline segment adds a new keyframe to the segment at the location where you clicked. With Step In or Step Out keyframes, clicking anywhere above or below a line segment moves that segment to where you’ve clicked. Time Stretch: If you select a range of keyframes, you can turn on the Time Stretch tool to show a box you can use to squeeze and stretch the entire range of keyframes relative to one another, to change the overall timing of a sequence of keyframes without losing the relative timing from one keyframe to the next. Alternately, you can turn on Time Stretch and draw a bounding box around the keyframes you want to adjust to create a time stretching boundary that way. Click Time Stretch a second time to turn it off. Shape Box: Turn on the Shape Box to draw a bounding box around a group of control points you want to adjust in order to horizontally squish and stretch (using the top/ bottom/left/right handles), cornerpin (using the corner handles), move (dragging on the box boundary), corner stretch (Command-drag the corner handles), Show Key Markers: Turning this control on shows keyframes in the top ruler that correspond to the frame at which each visible control point appears. The colors of these keyframes correspond to the color of the control points they’re indicating. Thumbnail Timeline Hidden by default, the Thumbnail timeline can be opened by clicking the Clips button in the UI Toolbar, and appears underneath the Node Editor when it’s open. The Thumbnail timeline shows you every clip in the current Timeline, giving you a way to navigate from one clip to another when working on multiple compositions in your project, and providing an interface for creating and switching among multiple versions of compositions, and resetting the current composition, when necessary. The Thumbnail timeline lets you navigate the timeline and manage versions of compositions Right-clicking on any thumbnail exposes a contextual menu The contextual menu for the thumbnail timeline Chapter – 9 Using the Fusion Page 2-37 To open another clip: Click any thumbnail to jump to that clip’s composition. The current clip is outlined in orange. To create and manage versions of compositions: To create a new version of a composition: Right-click the current thumbnail, and choose Create New Composition from the contextual menu. To load a different composition: Right-click the current thumbnail, and choose “NameOfVersion” > Load from the contextual menu. To delete a composition: Right-click the current thumbnail, and choose “NameOfVersion” > Delete from the contextual menu. To reset the current composition: Right-click the current thumbnail, and choose Reset Current Composition from the contextual menu. To change how thumbnails are identified: Double-click the area underneath any thumbnail to toggle among clip format, clip name, and a mystery that shall someday be solved by an intrepid team of adventurers embarking on a dangerous quest Media Pool In the Fusion page, the Media Pool continues to serve its purpose as the repository of all media you’ve imported into your project. This makes it easy to add additional clips to your compositions simply by dragging the clip you want from the Media Pool into the Node Editor. The media you add appears as a new MediaIn node in your composition, ready to be integrated into your node tree however you need. The Media Pool in Thumbnail mode showing video clips Chapter – 9 Using the Fusion Page 2-38 TIP: If you drag one or more clips from the Media Pool onto a connection line between two nodes in the Node Editor so that the connection highlights in blue and then drop them, those clips will be automatically connected to that line via enough Merge nodes to connect them all. The Bin List The Bin list at the left, which can be opened and closed, shows a hierarchical list of all bins used for organizing your media as well as your timelines. By default, the Media Pool consists of a single bin, named “Master,” but you can add more bins as necessary to organize timelines and clips by right-clicking anywhere in the empty area of the Media Pool and choosing Add Bin. You can rename any bin by double-clicking on its name and typing a new one, or by right-clicking a bin’s name and choosing Rename Bin. The Bin list can be hidden or shown via the button at the upper left-hand corner of the Fusion page toolbar. The Bin list The browser area to the right shows the contents of the currently selected bin in the Bin list. Every clip you’ve added, every timeline you’ve created, and every AAF, XML, or EDL file you’ve imported appears here. As elsewhere, the Media Pool can be displayed in either Icon or List view. In List view, you can sort the contents by any one of a subset of the total metadata that’s available in the Metadata Editor of the Media page. Of particular interest to audio editors are columns for Name, Reel Name, different timecode streams, Description, Comments, Keyword, Shot, Scene, Take, Angle, Circled, Start KeyKode, Flags, and Usage. For more information on using the myriad features of the Media Pool, see Chapter 8, “Adding and Organizing Media with the Media Pool.” In the sections that follow, some key features of the Media Pool are summarized for your convenience. Importing Media Into the Media Pool on the Fusion Page While adding clips to the Media Pool in the Media page provides the most organizational flexibility and features, if you find yourself in the Edit or Fairlight pages and you need to quickly import a few clips for immediate use, you can do so in a couple of different ways. Chapter – 9 Using the Fusion Page 2-39 To add media by dragging one or more clips from the Finder to the Fusion page Media Pool (macOS only): 1 Select one or more clips in the Finder. 2 Drag those clips into the Media Pool of DaVinci Resolve, or to a bin in the Bin list. Those clips are added to the Media Pool of your project. To use the Import Media command in the Fusion page Media Pool: 1 With the Fusion page open, right-click anywhere in the Media Pool, and choose Import Media. 2 Use the Import dialog to select one or more clips to import, and click Open. Those clips are added to the Media Pool of your project. For more information on importing media using the myriad features of the Media page, see Chapter 8, “Adding and Organizing Media with the Media Pool.” Bins, Power Bins, and Smart Bins There are actually three kinds of bins in the Media Pool, and each appears in its own section of the Bin list. The Power Bin and Smart Bin areas of the Bin list can be shown or hidden using commands in the View menu (View > Show Smart Bins, View > Show Power Bins). Here are the differences between the different kinds of bins: Bins: Simple, manually populated bins. Drag and drop anything you like into a bin, and that’s where it lives, until you decide to move it to another bin. Bins may be hierarchically organized, so you can create a Russian dolls nest of bins if you like. Creating new bins is as easy as right-clicking within the bin list and choosing Add Bin from the contextual menu. Power Bins: Hidden by default. These are also manually populated bins, but these bins are shared among all of the projects in your current database, making them ideal for shared title generators, graphics movies and stills, sound effects library files, music files, and other media that you want to be able to quickly and easily access from any project. To create a new Power Bin, show the Power Bins area of the Bin list, then rightclick within it and choose Add Bin. Smart Bins: These are procedurally populated bins, meaning that custom rules employing metadata are used to dynamically filter the contents of the Media Pool whenever you select a Smart Bin. Smart Bins are a a fast way of organizing the contents of projects for which you (or an assistant) has taken the time to add metadata to your clips using the Metadata Editor, adding Scene, Shot, and Take information, keywords, comments and description text, and myriad other pieces of information to make it faster to find what you’re looking for when you need it. To create a new Smart Bin, show the Smart Bin area of the Bin list (if necessary), then right-click within it and choose Add Smart Bin. A dialog appears in which you can edit the name of that bin and the rules it uses to filter clips, and click Create Smart Bin. Showing Bins in Separate Windows If you right-click a bin in the Bin List, you can choose “Open As New Window” to open that bin into its own window. Each window is its own Media Pool, complete with its own Bin list, Power Bins and Smart Bins lists, and display controls. Chapter – 9 Using the Fusion Page 2-40 This is most useful when you have two displays connected to your workstation, as you can drag these separate bins to the second display while DaVinci Resolve is in single screen mode. If you hide the Bin list, not only do you get more room for clips, but you also prevent accidentally switching bins if you really want to only view a particular bin’s contents in that window. You can as many additional Bin windows open as you care to, in addition to the main Media Pool that’s docked in the primary window interface. Filtering Bins Using Color Tags If you’re working on a project that has a lot of bins, you can apply color tags to identify particular bins with one of eight colors. Tagging bins is as easy as right-clicking any bin and choosing the color you want from the Color Tag submenu. For example, you can identify the bins that have clips you’re using most frequently with a red tag. A bin’s color tag then appears as a colored background behind that bin’s name. Using Color Tags to identify bins Once you’ve tagged one or more Media Pool bins, you can use the Color Tag Filter pop-up menu (the pop-up control to the right of the Bin List button) to filter out all but a single color of bin. Using Color Tag filtering to isolate the blue bins To go back to seeing all available bins, choose Show All from the Color Tag Filter pop-up. Chapter – 9 Using the Fusion Page 2-41 Sorting the Bin List The Bin list (and Smart Bin list) of the Media Pool can be sorted by bin Name, Date Created, or Date Modified, in either ascending or descending order. Simply right-click anywhere within the Bin list and choose the options you want from the Sort by submenu of the contextual menu. You can also choose User Sort from the same contextual menu, which lets you manually drag all bins in the Bin list to be in whatever order you like. As you drag bins in this mode, an orange line indicates the new position that bin will occupy when dropped. Dragging a bin to a new position in the Bin list in User Sort mode If you use User Sort in the Bin list to rearrange your bins manually, you can switch back and forth between any of the other sorting methods (Name, Date Created, Date Modified) and User Sort and your manual User Sort order will be remembered, making it easy to use whatever method of bin sorting is most useful at the time, without losing your customized bin organization. Searching for Content in the Media Pool An optional Search field can be opened at the top of the Media Pool that lets you quickly find clips by name, partial name, or any of a wide variety of Media Pool metadata. To search for a clip by name: 1 Select which bin or bins you want to search. 2 Click the magnifying glass button at the upper right-hand corner of the Media Pool. 3 Choose the particular column of information you want to search (or All Fields to search all columns) using the Filter by pop-up menu. Only selected bins will be searched. 4 Type your search string into the Search field that appears. A few letters should be enough to isolate only those clips that have that character string within their name. To show all clips again, click the cancel button at the right of the search field. TIP: Smart Bins are essentially multi-criteria search operations that scope the entire project at once and are saved for future use. Taking Advantage of the Media Pool’s Usage Column. In List view, the Usage column does not automatically update to show how many times a particular clip has been used. However, you can manually update this metadata by right-clicking within the Media Pool and choosing Update Usage Data from the contextual menu that appears. Afterwards, each clip will display how many times it’s been used in this column. Clips that have not been used yet display an x. Chapter – 9 Using the Fusion Page 2-42 The Console The Console, available by choosing Fusion > Console, is a window in which you can see the error, log, script, and input messages that may explain something the Fusion page is trying to do in greater detail. The Console is also where you can read FusionScript outputs, or input FusionScripts directly. Occasionally the Tooltip bar will display a badge to let you know there’s a message in the console you might be interested in. The badge will indicate if the message is an error, log, or script message. The Console window A toolbar at the top of the console contains controls governing what the console shows. At the top left, the Clear Screen button clears the contents of the Console. The next four buttons toggle the visibility of Error messages, Log messages, Script messages, and Input echoing. Showing only a particular kind of message can help you find what you’re looking for at when you’re under the gun at three in the morning. The next three buttons let you choose the input script language. Lua 5.1 is the default and is installed with Fusion. Python 2.7 and Python 3.3 require that you have the appropriate Python environment already installed on your computer. Since scripts in the Console are executed immediately, you can switch between input languages at any time. At the bottom of the Console is an Entry field. You can type scripting commands here for execution in the current comp context. Scripts are entered one line at a time, and are executed immediately. There are some useful shortcuts you can do in the Console. More information on scripting will be forthcoming as it becomes available. Chapter – 9 Using the Fusion Page 2-43 Customizing the Fusion Page This section explains how you can customize the Fusion page to accommodate whatever workflow you’re pursuing. The Fusion Settings Window The Fusion page has its own settings window, accessible by choosing Fusion > Fusion Settings. This window has a variety of options for customizing the Fusion experience, which will be documented in more detail at a later date. The Fusion Settings window set to the User Interface panel Showing and Hiding Panels The UI toolbar at the top of the screen lets you open panels you need and hide those you don’t. It’s the simplest way to create a layout for your particular needs at the moment. The UI toolbar of the Fusion page Chapter – 9 Using the Fusion Page 2-44 Resizing Panels You can change the overall size of each panel using preset configurations or you can adjust them manually. The Viewers and Work Panel are inverse of each other. The more space used to display the Work Panel, the less space available for the Viewers. To resize a panel, manually drag anywhere along the raised border surrounding the edges of the panel. Dragging the edge between two viewers to resize it Chapter – 9 Using the Fusion Page 2-45 Chapter 10 Getting Clips into the Fusion Page Chapter – 10 Getting Clips into the Fusion Page 2-46 Contents Getting Clips into the Fusion Page 2-48 Working on Single Clips in the Fusion Page 2-48 Adding Additional Media to Single-Clip Fusion Compositions 2-49 Adding Clips to Fusion From the File System 2-49 Creating Fusion Clips to Move Media Into the Fusion Page 2-50 Chapter – 10 Getting Clips into the Fusion Page 2-47 Getting Clips into the Fusion Page Now that Fusion compositing is integrated into DaVinci Resolve, it’s easy to get clips from your edit into the Fusion page to create any number of effects, prior to grading in the Color page. Depending on your needs, there are three ways that clips can find their way into the Fusion page. Working on Single Clips in the Fusion Page Each visible clip in a timeline appears in the Fusion page as a single MediaIn node connected to a MediaOut node. Clips that aren’t visible because they’re on lower tracks with fully opaque clips above them are ignored. These very-simple default compositions are referred to unofficially in this manual as “single-clip compositions.” The MediaIn node represents the image that’s fed to the Fusion page for further work, and the MediaOut node represents the final output that’s fed onward to the Color page for grading. The default node tree that appears when you first open the Fusion page while the playhead is parked on a clip This initial node structure makes it easy to quickly use the Fusion page to create relatively simple effects that are better accomplished using the power of node-based compositing. For example, if you have a clip that’s an establishing shot with no camera motion that needs some fast paint to cover up a bit of garbage in the background, you can open the Fusion page, add a Paint node, and use the Clone mode of the Stroke tool to quickly paint it out. A simple paint effect applied to a shot with no camera motion Once you’ve finished, simply go back to the Edit page and continue editing, because the entire Fusion composition is encapsulated within that clip, similarly to how grades in the Color page are also encapsulated within a clip. However you slip, slide, ripple, roll, or resize that clip, the Fusion effects you’ve created and the Color page grades you’ve made follow that clip’s journey through your edited timeline. TIP: While you’ll likely want to do all of the compositing for a green screen style effect in the Fusion page, it’s also possible to add a keyer, such as the exellent DeltaKeyer node, between the MediaIn and MediaOut nodes, all by itself. When you pull a key this way, the alpha channel is added to the MediaOut node, so your clip on the Edit page has transparency, letting you add a background clip on a lower track of your Edit page timeline. Chapter – 10 Getting Clips into the Fusion Page 2-48 Adding Additional Media to Single-Clip Fusion Compositions You’ll often find that even though you start out wanting to do something relatively simple to a single clip, you end up needing to add another clip or two to create the effect that you really need. For this reason, you can open the Media Pool on the Fusion page, and drag clips directly to the Node Editor to add them to your node tree. (Top) Dragging a clip from the Media Pool, (Bottom) Dropping it onto your composition When you do so by dragging a clip into an empty area of the Node Editor, the clip you dragged in becomes another MediaIn node, disconnected, and ready for you to merge into your current composite in any one of a variety of ways. TIP: If you drag a clip from the Media Pool directly on top of a connection line between any two other nodes in the Node Editor, that clip will automatically be added as the foreground clip connected to a Blend node that composites the new clip over the top of whatever you had before. When you add additional clips from the Media Pool, those clips becomes a part of the composition, similar to how Ext Matte nodes you add to the Color page Node Editor become part of that clip’s grade. Adding Clips to Fusion From the File System You also have the option of dragging clips from the file system directly into the Node Editor. When you do this, they’ll be added to the currently selected bin of the Media Pool automatically. So, if you have a library of stock animated background textures and you’ve just found one you want to use using your file system’s search tools, you can simply drag it straight into the Node Editor to use it right away. Chapter – 10 Getting Clips into the Fusion Page 2-49 Creating Fusion Clips to Move Media Into the Fusion Page For instances where you know you’re creating a more ambitious composited effect that requires multiple layers edited together with very specific timings, you can create a “Fusion clip” right from the Timeline. For example, if you have a foreground greenscreen clip, a background clip, and an additional graphic clip, then you can stack them all on the Timeline as superimposed clips, aligning their timings to work together as necessary by slipping, retiming, or otherwise positioning each clip. You can also edit multiple consecutive clips together that you want to use in a composition as a series of clips. Once that’s done, you can select every clip in the stack to create a Fusion clip, so you can easily use all these superimposed layers within a Fusion composite. To create a Fusion clip: 1 Edit all of the clips you want to use in the Edit page timeline. 2 Select all clips you want to be in the same composite at once. 3 Right-click one of the selected clips and choose New Fusion Clip from the contextual menu. 4 A new clip, named “Fusion Clip X” (where X is an automatically incrementing number) appears in the currently selected bin of the Media Pool and in the Timeline to replace the previously selected clips. 5 With the playhead parked over that clip, open the Fusion page to see the new arrangement of those clips in the Fusion page Node Editor. (Top) A stack of clips to use in a composite, (Bottom) Turning that stack into a Fusion clip in the Edit page Chapter – 10 Getting Clips into the Fusion Page 2-50 The nice thing about creating a Fusion clip is that every superimposed clip in a stack is automatically connected together into a cascading series of Merge nodes that creates the desired arrangement of clips. Note that whatever clips were in the bottom of the stack in the Edit page appear at the top of the Node Editor in the Fusion page, but the arrangement of background and foreground input connections is appropriate to recreate the same compositional order. The initial node tree of the three clips we turned into a Fusion clip Chapter – 10 Getting Clips into the Fusion Page 2-51 Chapter 11 Image Processing and Color Management This chapter covers how the Fusion page fits into the overall image processing pipeline of DaVinci Resolve 15. It also discusses the value of doing compositing with clips using a linear gamma, and how to deal with color management in the Fusion page, so you can work with a linear gamma while previewing the image in the Viewer using the gamma of your choice. Chapter – 11 Image Processing and Color Management 2-52 Contents Fusion’s Place in the Resolve Image Processing Pipeline 2-54 Source Media into the Fusion Page 2-54 Edit Page Plug-ins and the Fusion Page 2-54 Forcing Effects into the Fusion Page by Making Compound Clips 2-54 Output from the Fusion page to the Color page 2-55 What Viewers Show in Different Pages 2-55 Sizing and the Fusion Page 2-55 Color Management and the Public Beta 2-55 Converting to Linear in the Fusion Page 2-56 Viewer Gamma and Color Space While Working in Linear 2-56 Chapter – 11 Image Processing and Color Management 2-53 Fusion’s Place in the Resolve Image Processing Pipeline In most workflows, clips are are edited, then effects are applied to the edited clips, and the clips with these effects are graded, in pretty much that order. This is the order of operations that DaVinci Resolve follows, source clips edited into the Timeline in the Edit page flow into the Fusion page node tree, and image data from the Fusion page node tree flows into the Color page. DaVinci Resolve goes so far as to expose this via the order of the page buttons at the bottom of the screen, with the Edit page feeding the Fusion page, and the Fusion page feeding the Color page. However, this isn’t the whole story. As of the public beta of DaVinci Resolve 15, the following sections describe which effects happen prior to the Fusion page, and which effects happen after the Fusion page. NOTE: Since this documentation covers the public beta of DaVinci Resolve 15, this information may change as new versions become available. Source Media into the Fusion Page For ordinary clips, the MediaIn nodes in the Fusion page represent each clip’s source media, as modified by whatever changes you’ve imposed on the source media via the Clip Attributes window, and whatever Edit page Transform and Cropping adjustments you’ve made to that clip. Edit Page Plug-ins and the Fusion Page If you add a ResolveFX or an OFX plug-in to a clip in the Edit page, and then you open the Fusion page, you won’t see that plug-in taking effect. That’s because these plug-ins are actually applied after the output of the Fusion page, but before the input of the Color page. If you open the Color page, you’ll see the Edit page plug-in being applied to that clip, effectively as an operation prior to the grading adjustments and effects you apply in the Color page Node Editor. With this in mind, the order of effects processing in the different pages of DaVinci Resolve can be described as follows: Source Media > Clip Attributes > Edit Sizing > Fusion Effects > Edit Plug-ins (ResolveFX) > Color Effects Forcing Effects into the Fusion Page by Making Compound Clips There is a way you can force clips with Edit page ResolveFX and OFX and Color page grades into the Fusion page, and that is to turn that clip into a compound clip. When Edit page effects and Color page grading are embedded within compound clips, MediaIn nodes corresponding to compound clips route the effected clip into the Fusion page. Chapter – 11 Image Processing and Color Management 2-54 Output from the Fusion page to the Color page The composition output by the Fusion page’s MediaOut node are propagated via the Color page’s source input, with the sole exception that if you’ve added plug-ins to that clip in the Edit page, then the handoff from the Fusion page to the Color page is as follows: Fusion Effects > Edit page Plug-ins > Color Effects What Viewers Show in Different Pages Owing to the different needs of compositing artists, editors, and colorists, the Viewers in the public beta show different states of the clip. The Edit page Source Viewer: Always shows the source media. If Resolve Color Management is enabled, then the Edit page Source Viewer will show the source media at the Timeline color space and gamma. The Edit page Timeline Viewer: Shows clips with all Edit page effects, Color page grades, and Fusion page effects applied, so editors see the program within the context of all effects and grading. The Fusion page Viewer: Shows clips at the Timeline color space and gamma, with no Edit page effects and no Color page grades. The Color page Viewer: Shows clips with all Edit page effects, Color page grades, and Fusion page effects applied. Sizing and the Fusion Page With the addition of the Fusion page, the order of sizing operations in DaVinci Resolve is a bit more complex. However, it’s important to understand which sizing operations happen prior to the Fusion page, and which happen after, so you know which effects will alter the image that’s input to the Fusion page, and which effects happen to the Fusion page’s output. For example, Lens Correction, while not strictly sizing, is nonetheless an effect that will change how the image begins in your Fusion composition. However, Stabilization is an effect that comes after the Fusion page, so it has no effect on the composition you’re creating. The order of sizing effects in the different pages of DaVinci Resolve can be described as follows: Super Scale > Edit Sizing/Lens Correction > Fusion Transforms > Stabilization > Input Sizing > Output Sizing Color Management and the Public Beta At the time of this writing, the Fusion page does not automatically interact in any way with Resolve Color Management (RCM). Images coming into the Fusion page via MediaIn nodes are in the Timeline gamma and color space. For some simple operations, this may be fine, but it’s not always ideal. Chapter – 11 Image Processing and Color Management 2-55 Converting to Linear in the Fusion Page Because node operations in the Fusion page handle image data in very direct ways, you should ideally composite images that use a linear gamma, especially when you’re combining images and effects involving bright highlights. This is because common operations such as operations that divide an image (aka “unpremultiply”), composite modes such as “screen,” merge operations, and many other compositing tasks only work properly with a linear gamma. For example, you can apply filtering effects, such as a blur, to an image using any gamma setting, and the image will probably look fine. However, if you convert the image to a linear gamma first and then apply the blur, then images (especially those with extremely bright areas) will be processed with greater accuracy, and you should notice a different and superior result. Fortunately, the Fusion page has manual tools that let you convert each MediaIn clip from the timeline gamma to linear gamma at the beginning of your composite, and then convert from linear back to the timeline gamma at the end of your composite, right before the MediaOut node feeds its result to the Color page. The CineonLog node, found in the Film category of the Effects Library, lets you do a conversion from any of the formats in the Log Type pop-up menu to Linear, and vice versa. This is useful if your timeline gamma is a Log format. The FileLUT node, found in the LUT category of the Effects Library, lets you do a conversion using any LUT you want, giving you the option to manually load one of the LUTs that accompany DaVinci Resolve in the /Library/Application Support/Blackmagic Design/DaVinci Resolve/LUT/VFX IO/ directory (on macOS) to perform to/from linear gamma conversions. A node tree with “to linear” conversions at the beginning, and a “from linear” conversion at the end NOTE: In the standalone version of Fusion, “Loader” nodes have color space and gamma conversion built-in when you expose their controls in the Inspector. However, this functionality has not yet been added to the public beta of DaVinci Resolve 15. Viewer Gamma and Color Space While Working in Linear Images converted to a linear gamma don’t look correct. In fact they usually look terrible. Since all image data is converted to a linear scale for the convenience and accuracy of compositing operations, highlights usually get stretched to look extremely bright and blown out, and colors can become exaggerated and oversaturated. Happily, even though the image looks incorrect, the fact that DaVinci Resolve works entirely with 32-float color data internally means that, despite how this looks, you’re not clipping or losing any image data. It just looks bad when viewing the naked state of your image data. Chapter – 11 Image Processing and Color Management 2-56 It would be obviously impossible to work if you couldn’t see the image as it’s supposed to look, in the final gamma you’ll be outputting to. For this reason, each Viewer has a LUT control that lets you enable a “preview” color space and/or gamma conversion that lets you see the image in your intended color space and gamma, while the node tree is processing correctly in linear gamma. The Viewer LUT button turned on, choosing a LUT to use from the pop-up menu Clicking the Viewer LUT button toggles LUT display on or off, while its accompanying pop-up menu lets you choose which of the many available color space and gamma conversions to view with. The top five options let you choose Fusion controls, which can be customized via the Edit item at the bottom of this menu. The rest of this menu shows all LUTs from the /Library/ Application Support/Blackmagic Design/DaVinci Resolve/LUT/VFX IO/ directory (on macOS) to use for viewing. So, if you’re working in linear, you can choose VFX IO > Linear to Gamma 2.4 to see a “normalized” version of the composite you’re working on. NOTE: By default, Viewers in Fusion show you the image prior to any grading done in the Color page, since the Fusion page comes before the Color page in the DaVinci Resolve image processing pipeline. Chapter – 11 Image Processing and Color Management 2-57 Chapter 12 Understanding Image Channels and Node Processing Chapter – 12 Understanding Image Channels and Node Processing 2-58 Contents Channels in the Fusion Page 2-60 Types of Channels Supported by Fusion 2-60 Fusion Node Connections Carry Multiple Channels 2-61 Node Inputs and Outputs 2-61 Node Colors Tell You Which Nodes Connect 2-65 Using Channels in a Composition 2-67 Channel Limiting 2-69 Adding Alpha Channels 2-69 How Channels Propagate During Compositing 2-70 Rearranging or Combining Channels 2-71 Understanding Premultiplication 2-72 The Rules of Premultiplication 2-73 How Do You Know You’ve Made a Premultiplication Mistake? 2-73 Setting the Premultiplied Status of MediaIn Nodes That Need It 2-73 Nodes That Affect Premultiplication 2-73 Control Premultiplication with Alpha Divide and Alpha Multiply 2-74 Understanding Auxiliary Channels 2-74 Image Formats That Support Auxiliary Channels 2-75 Creating Auxiliary Channels in Fusion 2-75 Auxiliary Channels Explained 2-76 Propagating Auxiliary Channels 2-80 Viewing Auxiliary Channels 2-80 Nodes That Use Auxiliary Channels 2-81 Merge 2-81 Depth Blur 2-81 Fog 2-81 Shader 2-81 SSAO 2-81 Texture 2-81 Shadow 2-81 Vector Motion Blur 2-82 Vector Distortion 2-82 Time Speed and Time Stretcher 2-82 New Eye 2-82 Stereo Align 2-82 Smooth Motion 2-82 Volume Fog 2-82 Volume Mask 2-82 Lumakeyer 2-82 Copy Aux 2-83 Channel Boolean 2-83 Chapter – 12 Understanding Image Channels and Node Processing 2-59 Channels in the Fusion Page If you’re an old hand at compositing in Fusion, this chapter may be somewhat remedial. However, the Fusion page introduces some innovative ways of working with the many different channels of image data that modern compositing workflows encompass. In particular, many shortcuts for handling different kinds of channels have been built into the way that different nodes interact with one another, making this chapter’s introduction to color channels and how they’re affected by different nodes and operations a valuable way to begin the process of learning to do paint, compositing, and effects in the Fusion page. If you’re new to compositing, or you’re new to the Fusion workflow, you ignore this chapter at your peril, as it provides a solid foundation to understanding how to predictably control image data as you work in this powerful environment. Types of Channels Supported by Fusion Digital images are divided into separate channels, each of which carries a specific kind of image data. Nodes that perform different image processing operations typically expect specific channels in order to provide predictable results. This section describes the different kinds of channels that the Fusion page supports. Incidentally, all image data in DaVinci Resolve, including the Fusion page, is 32-bit float. RGB Color channels The Red, Green, and Blue channels of any still image or movie clip combine additively to represent everything we can see via visible light. Each of these three channels is a grayscale image when seen by itself, as seen in the following screenshots. When combined additively, these channels represent a full-color image. Alpha Channels An Alpha channel is a grayscale channel that represents different levels of transparency in an image. In Fusion, white denotes areas that are solid, while black denotes areas that are transparent. Grayscale values range from more opaque (lighter) to more transparent (darker). If you’re working with an imported Alpha channel from another application for which these conventions are reversed, never fear. Every node capable of using an Alpha channel is also capable of inverting it. Single-Channel Masks These channels are created by Fusion whenever you create a Mask node. Mask nodes are unique in that they propagate single-channel image data that often serves a similar function as an alpha channel, defining which areas of an image should be solid and which should be transparent. However, Masks can also define which parts of an image should be affected by a particular operation, and which should not. Mask channels are designed to be connected to specific mask inputs of nodes used for keying and compositing, such as the Merge node, the DeltaKeyer node, and the Matte Control node. In the following example, you can see how a Mask can be used as a garbage matte for cropping unwanted background equipment out of a greenscreen layer. Auxiliary Channels Auxiliary channels (covered in more detail later in this chapter), describe a family of special‑purpose image data that typically expose 3D data in a way that can be used in 2D composites. For example, Z-Depth channels describe the depth of each feature in an Chapter – 12 Understanding Image Channels and Node Processing 2-60 image along a Z axis (XYZ), while an XYZ Normals channel describes the orientation (facing up, facing down, or facing to the left or right) of each pixel in an image. Auxiliary channel data is generated by rendering 3D images and animation, so it usually accompanies images generated by Autodesk Maya or 3DS Max, or it may be generated from within the Fusion page via the Renderer 3D node, which outputs a 3D scene that you’ve assembled and lit as 2D RGBA channels, with optionally accompanying Auxiliary channels. The reason to use Auxiliary data is that 3D rendering is computationally expensive and time‑consuming, so outputting descriptive information about a 3D image that’s been rendered empowers compositing artists to make sophisticated alterations in 2D to fine-tune focus, lighting, and depth compositing that are faster (cheaper) to perform and readjust in 2D than re-rendering the 3D source material over and over. TIP: You can view any of a node’s channels in isolation using the Color control in the Viewer. Clicking the Color control switches between Color (RGB) and Alpha, but clicking its pop-up menu control reveals a list of all channels within the currently selected node, including red, green, blue, or auxiliary channels. Fusion Node Connections Carry Multiple Channels The connections that pass image data from one node to the next in the Node Editor of the Fusion page are capable of carrying multiple channels of image data along a single line. That means that a single connection may route RGB, or RGBA, or RGBAZ-Depth, or even just Z-Depth, depending on how you’ve wired up your node tree. In the following example, each of the two MediaIn nodes output RGB data. However, the Delta Keyer adds an Alpha channel to the foreground image that the Merge node can use to create a two-layer composite. MediaIn2 node connected to a DeltaKeyer node, connected to a Merge node, which is connected to another MediaIn node to combine the two images using the alpha channel output by the DeltaKeyer Running multiple channels through single connection lines makes Fusion node trees simple to read, but it also means you need to keep track of which nodes process which channels to make sure that you’re directing the intended image data to the correct operations. Node Inputs and Outputs MediaIn nodes output all available channels from the source media on disk. When you connect one node’s output to another node’s input, those channels are passed from the upstream node to the downstream node, which then processes the image according to that node’s function. Only one node output can be connected to a node input at a time. In this simple example, a MediaIn node’s output is connected to the input of a Hilight node to create a sparkly highlight effect. Chapter – 12 Understanding Image Channels and Node Processing 2-61 MediaIn node connected to a Highlight node connected to MediaOut node When connecting nodes together, a single node output can be connected to multiple node’s inputs, which is known as “branching.” This is useful when you have a single node that you want to feed multiple operations at the same time. The MediaIn node’s output is branched to the inputs of two other nodes Using Multiple Inputs Most nodes have two inputs, one for RGBA and another for a mask that can be optionally used to limit the effect of that node to a particular part of the image (a similar idea to using a KEY input to perform secondary corrections in the Color page). However, some nodes have three or even more inputs, and it’s important to make sure you connect the correct image data to the appropriate input in order to obtain the desired result. If you connect a node to another node’s input and nothing happens, chances are you’ve connected to the wrong input. For example, the MatteControl node has a background input and a foreground input, both of which accept RGBA channels. However, it also has SolidMatte, GarbageMatte, and EffectsMask inputs that accept Matte or Mask channels to modify the alpha key being extracted from the image in different ways. If you want to perform the extremely common operation of using a MatteControl node to attach a Polygon node for rotoscoping an image, you need to make sure that you connect the Polygon node to the GarbageMatte input to obtain the correct result, since the GarbageMatte input is automatically set up to use the input mask to alter the alpha channel of the image. If you connect to any other input, your Polygon mask won’t work. Polygon node connected to a MatteControl node for rotoscoping Chapter – 12 Understanding Image Channels and Node Processing 2-62 In another example, the DeltaKeyer node has a primary input (labeled “Input”) that accepts RGBA channels, but it also has a CleanPlate input for attaching an RGB image with which to clean up the background (typically the CleanPlate node), and SolidMatte, GarbageMatte, and EffectsMask inputs that accept Matte or Mask channels to modify the alpha key being extracted from the image in different ways. To pull a key successfully, though, you must connect the image you want to key to the “Input” input. MediaIn node connected to the main “Input” input of a DeltaKeyer node, other nodes connected to specific inputs for those particular nodes If you position your pointer over any node’s input or output, a tooltip will appear in the Tooltip bar at the bottom of the Fusion page letting you know what that input or output is for, to help guide you to using the right input for the job. If you pause for a moment longer, another tooltip will appear in the Node Editor itself. (Left) The node input’s tooltip in the Tooltip bar, (Right) The node tooltip in the Node Editor Connecting to the Correct Input When you’re connecting nodes together, pulling a connection line from the output of one node and dropping it right on top of the body of another node makes a connection to the default input for that node, which is commonly “Input” or “Background.” Side by side, dropping a connection on a node’s body to connect to that node’s primary input Chapter – 12 Understanding Image Channels and Node Processing 2-63 However, if you drop a connection line right on top of a specific input, then you’ll connect to that input, so it’s important to be mindful of where you drop connection lines as you wire up different node trees together. Side by side, dropping a connection on a specific node input, note how the inputs rearrange themselves afterwards to keep the node tree tidy-looking TIP: If you hold the Option key down while you drag a connection line from one node onto another, and you keep the Option key held down while you release the pointer’s button to drop the connection, a menu appears that lets you choose which specific input you want to connect to, by name. Some Nodes are Incompatible With Some Inputs Usually, you’re prevented from connecting a node’s output to another node or node input that’s not compatible with it. For example, if you try to connect a Merge3D node’s output directly to the input of a regular Merge node, it won’t work; you must first connect to a Renderer3D node that creates output appropriate for 2D compositing operations. In other cases, connecting the wrong image data to the wrong node input won’t give you any sort of error, it will simply fail to produce the result you were expecting, necessitating you to troubleshoot the composition. If this happens to you, check the Node Reference section (or the Fusion Tool Manual for previous versions of Fusion) to see if the node you’re trying to connect to has any limitations as to how it must be attached. TIP: This chapter tries to cover many of the little exceptions to node connection that are important for you to know, so don’t skim too fast. Always Connect the Background Input First Many nodes combine images in different ways using “background” and “foreground” inputs, including the Merge node, the Matte Control node, and the Channel Booleans node as commonly used examples. To help you keep things straight, background inputs are always orange, and foreground inputs are always green. Chapter – 12 Understanding Image Channels and Node Processing 2-64 When you first connect any node’s output to a multi-input node, you usually want to connect the background input first. This is handled for you automatically when you first drop a connection line onto the body of a new multi-input node, it usually connects to the orange-colored background input first (the exception is mask nodes, which always connect to the first available mask input). This is good, because you want to get into the habit of always connecting the background input first. If you connect to only one input of a multi-input node and you don’t connect to the background input, you may find that you don’t get the results you wanted. This is because each multi-input node expects that the background will be connected before anything else, so that the internal connections and math used by that node can be predictable. TIP: The only node to which you can safely connect the foreground input prior to the background input is the Dissolve node, which is a special node that can be used to either dissolve between two inputs, or automatically switch between two inputs of unequal duration. Node Colors Tell You Which Nodes Connect Unlike the Color page, where each Corrector node is capable of performing nearly every kind of grading operation in combination for speed of grading, each of the many nodes in the Fusion page accomplish a single type of effect or operation. These single-purpose nodes make it easier to decipher a complex composition when examining its node tree, and it also makes it easier for artists to focus on fine-tuning specific adjustments, one at a time, when assembling the ever-growing tree of MediaIn nodes and image processing operations that comprise one’s composite. Because each Fusion page node has a specific function, they’re categorized by type to make it easier to keep track of which nodes require what types of image channels as input, and what image data you can expect each node to output. These general types are described here. A node tree showing the main categories of node colors Chapter – 12 Understanding Image Channels and Node Processing 2-65 Blue MediaIn Nodes and Green Generator Nodes Blue MediaIn nodes add clips to a composite, and green Generator nodes create images. Both types of nodes output RGBA channels (depending on the source and generator), and may optionally output auxiliary channels for doing advanced compositing operations. Because these are sources of images, both kinds of nodes can be attached to a wide variety of other nodes for effects creation besides just 2D nodes. For example, you can also connect MediaIn nodes to Image Plane 3D nodes for 3D compositing, or to pEmitter nodes set to “Bitmap” for creating different particle systems. Green Generator nodes can be similarly attached to many different kinds of nodes, for example attaching a FastNoise node to a Displace 3D node to impose undulating effects to 3D shapes. 2D Processing Nodes, Color Coded by Type These encompass most 2D processing and compositing operations in DaVinci Resolve, all of which process RGBA channels and pass along auxiliary channels. These include: Orange Blur nodes Olive Color Adjustment nodes (color adjustment nodes additionally concatenate with one another) Pink Paint nodes Dark orange Tracking nodes Tan Transform node (transform nodes additionally concatenate with one another) Teal VR nodes Dark brown Warp nodes Gray which includes Compositing nodes as well as many other types. Additionally, some 2D nodes such as Fog and Depth Blur (in the Deep Pixel category) accept and use Auxiliary channels such as Z-Depth to create different perspective effects in 2D. TIP: Two 2D nodes that specifically don’t process alpha channel data are the Color Corrector node, designed to let you color correct a foreground layer to match a background layer without affecting an alpha channel being used to create a composite. and the Gamut node, which lets you perform color space conversions to RGB data from one gamut to another without affecting the alpha channel. Purple Particle System Nodes These are nodes that connect together to create different particle systems, and they’re incompatible with other kinds of nodes until you add a pRender node which outputs 2D RGBA and Auxiliary data that can be composited with other 2D nodes and operations. Dark Blue 3D Nodes These are 3D operations, which generate and manipulate 3D data (including auxiliary channels) that is incompatible with other kinds of nodes until processed via a Renderer 3D node, which then outputs RGBA and Auxiliary data. Chapter – 12 Understanding Image Channels and Node Processing 2-66 Brown Mask Nodes Masks output single-channel images that can only be connected to one another (to combine masks) or to specified Mask inputs. Masks are useful for defining transparency (Alpha masks), defining which parts of an image should be cropped out (Garbage masks), or defining which parts of an image should be affected by a particular node operation (Effects masks). Using Channels in a Composition When you connect one node’s Output to another node’s Input, you feed all of the channels that are output from the upstream node to the downstream node. 2D nodes, which constitute most simple image processing operations in the Fusion page, propagate all channel data from node to node, including RGB, alpha, and auxiliary channels, regardless of whether or not that node actually uses or affects a particular channel. Incidentally, if you want to see which channels are available for a node, you can open the Color pop-up menu in the Viewer to get a list. This control also lets you view any channel on this list, so you can examine the channel data of your composite anywhere along the node tree. All channels that are available to the currently viewed node can be isolated via the Viewer’s Color control 2D nodes also typically operate upon all channel data routed through that node. For example, if you connect a node’s output with RGBA and XYZ Normals channels to the input of a Vortex node, all channels are equally transformed by the Size, Center, and Angle parameters of this operation, including the alpha and XYZ normals channels, as seen in the following screenshot. Chapter – 12 Understanding Image Channels and Node Processing 2-67 (Left) The Normal Z channel output by a rendered torus, (Right) The Normal Z channel after the output is connected to a Vortex node, note how this auxiliary channel warps along with the RGB and A channels This is appropriate, because in most cases you want to make sure that all channels are transformed, warped, or adjusted together. You wouldn’t want to shrink the image without also shrinking the alpha channel along with it, and the same is true for most other operations. On the other hand, some nodes deliberately ignore specific channels, when it makes sense. For example, the Color Corrector and Gamut nodes, both of which are designed to alter RGB data specifically, have no effect on alpha or auxiliary channels. This makes them convenient for color-matching foreground and background layers you’re compositing, without worrying that you’re altering the transparency or depth information accompanying that layer. MediaIn, DeltaKeyer, Color Corrector, and Merge/MediaIn node TIP: If you’re doing something exotic and you actually want to operate on a channel that’s usually unaffected by a particular node, you can always use the Channel Booleans node to reassign the channel you want to modify to another output channel that’s compatible with the operation you’re trying to perform, and then use another Channel Booleans node to reassign it back. When doing this to a single image, it’s important to connect that image to the background input of the Channel Booleans node, so the alpha and auxiliary channels are properly handled. Chapter – 12 Understanding Image Channels and Node Processing 2-68 Channel Limiting Most nodes have a set of Red, Green, Blue, and Alpha checkboxes in the Settings panel of that node’s controls in the Inspector. These checkboxes let you exclude any combination of these channels from being processed by that node. The channel limiting checkboxes in the Settings panel of a Transform node so only the Green channel is affected For example, if you wanted to use the Transform node to add a bump map to only the green channel of an image, you can turn off the Green, Blue, and Alpha checkboxes. As a result, the green channel is processed by this operation, and the red, blue, and alpha channels are copied straight from the node’s input to the node’s output, skipping that node’s processing to remain unaffected. Transforming only the Green color channel of the image with a Transform effect Adding Alpha Channels One of the whole reasons for compositing is to begin with an image that lacks an alpha channel, add one via keying or rotoscoping, and then composite that result against other images. While the methods for this are covered in detail in later chapters, here’s an overview of how this is handled within the Fusion page. In the case of extracting a alpha matte from a greenscreen image, you typically connect the image’s RGB output to the “Input” input of a Keyer node such as the Delta Keyer, and you then use the keyer’s controls to pull the matte. The Keyer node automatically inserts the alpha channel that’s generated alongside the RGB channels, so the output is automatically RGBA. Then, when you connect the Keyer’s output to a Merge node in order to composite it over another image, the Merge node automatically knows to use the embedded alpha channel coming into the foreground input to create the desired composite, as seen in the following screenshot. Chapter – 12 Understanding Image Channels and Node Processing 2-69 A simple node tree for keying, note that only one connection links the DeltaKeyer to the Merge node In the case of rotoscoping using a Polygon node, you’ll typically connect the image being rotoscoped to the background input of a MatteControl node, and a Polygon node to its Garbage Matte input (which you invert in the Inspector, unless you invert the Polygon’s output). This lets you view the image while drawing, using the controls of the Polygon node, and the resulting alpha channel is merged together with the RGB channels so the Merge Alpha node’s output is RGBA, which can be connected to a Merge node to composite the rotoscoped subject over another image. A simple rotoscoping node tree In both cases, you can see how the Fusion page node tree’s ability to carry multiple channels of image data over a single connection line simplifies the compositing process. How Channels Propagate During Compositing As you’ve seen above, images are combined, or composited together, using the Merge node. The Merge node takes two RGBA inputs labeled “Foreground” (green) and “Background” (orange) and combines them into a single RGB output (or RGBA if both the foreground and background input images have alpha), where the foreground image is in front (or on top, depending on what you’re working on), and the background image is, you guessed it, in back. A simple Merge node composite Chapter – 12 Understanding Image Channels and Node Processing 2-70 Auxiliary channels, on the other hand, are handled in a much more specific way. When you composite two image layers using the Merge node, auxiliary channels will only propagate through the image that’s connected to the background input. The rationale for this is that in most composites that include computer generated imagery, the background is most often the CG layer that contains auxiliary channels, while the foreground is a live-action greenscreen plate with subjects or elements that are meant to be combined against the background. Many compositions use multiple Merge nodes to bring together many differently processed branches of a large node tree, so it pays to be careful about how you connect the background and foreground inputs of each Merge node to make sure that the correct channels flow properly. TIP: Merge nodes are also capable of combining the foreground and background inputs using Z-Depth channels using the “Perform Depth Merge” checkbox, in which case every pair of pixels are compared. Which one is in front depends on its Z-Depth and not which input it’s connected to. Rearranging or Combining Channels Last, but certainly not least, it’s also possible to rearrange and re-combine channels in any way you need, using one of three different node operations. For example, you might want to combine the red channel from one image with the blue and green channels of a second image to create a completely different channel mix. Alternately, you might want to take the Alpha channel from one image and merge it with the alpha channel of a second image in different ways, adding, subtracting, or using other intersection operations to create a very specific blend of the two. The following nodes are used to re-combine channels in different ways: Channel Boolean: Used to switch among or combine two sets of input channels in different ways, using a variety of simple pre-defined imaging math operations. Channel Booleans: Used to rearrange YRGB/auxiliary channels within a single input image, or among two input images, to create a single output image. If you only connect a single image to this node, it must be connected to the background input to make sure everything works. Matte Control: Designed to do any combination of the following: (a) recombining mattes, masks, and alpha channels in various ways, (b) modifying alpha channels using dedicated matte controls, and (c) copying alpha channels into the RGB stream of the image connected to the background input in preparation for compositing. You can copy specific channels from the foreground input to the background input to use as an alpha channel, or you can attach masks to the garbage matte input to use as alpha channels as well. Chapter – 12 Understanding Image Channels and Node Processing 2-71 Understanding Premultiplication Now that you understand how to direct and recombine image, alpha, and auxiliary channels in the Fusion page, it’s time to learn a little something about premultiplication, to make sure you always combine RGB and alpha channels correctly to get the best results from Merge node composites. Premultiplication is an issue whenever you find yourself compositing multiple images together, and at least one of them contains RGB with an Alpha channel. For example, if a motion graphics artist gives you a media file with an animated title graphic that has transparency rendered into it to accommodate later compositing, or if an animator gives you an isolated VFX plate of a spaceship coming in for a landing with the transparency baked in, you may need to consider the pre-multiplied state of the RGBA image data as you use these images. Most computer-generated images you’ll be given should be premultiplied. A premultiplied alpha channel means that, for every pixel of an image, the RGB channels are multiplied by the alpha channel. This is standard practice in VFX workflows, and it guarantees that translucent parts of the rendered image, such as flares, smoke, or atmospheric effects, are correctly integrated into the background black areas of the isolated image, so that the image appears correct when you view that layer by itself. NOTE: Computer generated 3D images that were rendered anti-aliased are almost always premultiplied. So-called “straight” alpha channels, where the RGB channels have not been multiplied by the alpha channel, will appear weirdly bright in these same translucent areas, which tells you that you probably need to multiply the RGB and A channels prior to doing specific tasks. (Top) Premultiplied alpha image, (Bottom) Straight alpha image Chapter – 12 Understanding Image Channels and Node Processing 2-72 The Rules of Premultiplication In general, when you’re compositing multiple images together, and one or more has a built-in alpha channel, you want to make sure you follow these general rules: Always color-correct images that are not premultiplied Always filter and transform images that are premultiplied How Do You Know You’ve Made a Premultiplication Mistake? Improper handling of premultiplication manifests itself in two obvious ways: You see thin fringing around a subject composited with a Merge node You notice a node adjustment affecting parts of the image that shouldn’t be affected by that operation You’ve combined RGB and alpha channels from different sources and the checkerboard background pattern in the Viewer (if enabled) is only semi-transparent when it should be fully transparent If you spot these sorts of issues, the good news is they’re easy to fix using either the internal settings of the nodes causing the problem, or with dedicated nodes to force the premultiplied state of the image at specific points in your node tree. Setting the Premultiplied Status of MediaIn Nodes That Need It When you select a MediaIn node, the Import panel in the Inspector have a group of checkboxes that let you determine how an alpha channel embedded with that image should be handled. There are checkboxes to Make the alpha channel solid (to eliminate transparency), to invert the alpha channel, and to Post-Multiply the RGB channels with the alpha channel, should that be necessary. NOTE: This functionality was not yet available in the Public Beta of DaVinci Resolve 15 as of this writing. Nodes That Affect Premultiplication Most nodes that require you to explicitly deal with the state of premultiplication of RGBA image input have a “Pre-Divide, Post-Multiply” checkbox. This includes simple color correction nodes such as Brightness Contrast and Color Curves, as well as the Color Correct node, which has the “Pre-Divide/Post-Multiply” checkbox in the Options panel of its Inspector settings. The Pre-Divide/Post-Multiply checkbox of the Color Curves node, seen in the Inspector Chapter – 12 Understanding Image Channels and Node Processing 2-73 Control Premultiplication with Alpha Divide and Alpha Multiply The Alpha Divide and Alpha Multiply nodes, found in the Matte category of the Effects Library, are provided whenever you need to do operations on RGBA image data where you need explicit control over the pre-multiplied state of an image’s RGB channels against its alpha channel. Simply add the Alpha Divide node when you want the RGBA image data to not be premultiplied, and add the Alpha Multiply node when you want the image data to be premultiplied again. For example, if you’re using third-party OFX nodes that make color adjustments, you may need to manually control premultiplication before and after such an adjustment. A node tree with explicit Alpha Divide and Alpha Multiply nodes Understanding Auxiliary Channels Auxiliary channels describe a family of special-purpose image data that typically describes 3D position, orientation, and object information for use in 2D composites. For example, Z-Depth channels describe the depth of each region of an image along a Z axis (XYZ), while an XYZ Normals channel describes the orientation (facing up, facing down, facing to the left or right) of each pixel in an image. Auxiliary channel data is generated by rendering 3D data, so it may accompany images generated by Autodesk Maya or 3DS Max, or it may be generated from within the Fusion page via the Renderer 3D node, which outputs a 3D scene that you’ve assembled and lit as 2D RGBA channels, with optionally accompanying Auxiliary channels. One of the most common reasons to use Auxiliary data is that 3D rendering is computationally expensive and time-consuming, so outputting descriptive information about a 3D image that’s been rendered empowers compositing artists to make sophisticated alterations in 2D affecting focus, lighting, and depth compositing, that are faster to perform and readjust in 2D it would be to re-render the 3D source material over and over. There are two ways of obtaining Auxiliary channel data: First, auxiliary data may be embedded within a clip exported from a 3D application that’s in a format capable of containing Auxiliary channels. In this case, it’s best to consult your 3D application’s documentation to determine which auxiliary channels can be generated and output. You may also obtain Auxiliary channel data by generating it within the Fusion page, via 3D operations output by the Renderer 3D node, using the Optical Flow node, or using the Disparity node. Chapter – 12 Understanding Image Channels and Node Processing 2-74 An RGBA 3D rendered scene that can also generate auxiliary channels Image Formats That Support Auxiliary Channels Fusion supports auxiliary channel information contained in a variety of image formats. The number of channels and methods used are different for each format. OpenEXR (*.exr) The OpenEXR file format can contain an arbitrary number of additional image channels. Many renderers that will write to the OpenEXR format will allow the creation of channels that contain entirely arbitrary data. For example, a channel with specular highlights might exist in an OpenEXR. In most cases, the channel will have a custom name that can be used to map the extra channel to one of the channels recognized by Fusion. SoftImage PIC (*.PIC, *.ZPIC and *.Z) The PIC image format (used by SoftImage) can contain Z-Depth data in a separate file marked by the ZPIC file extension. These files must be located in the same directory as the RGBA PIC files and must use the same names. Fusion will automatically detect the presence of the additional information and load the ZPIC images along with the PIC images. Wavefront RLA (*.RLA), 3ds Max RLA (*.RLA) and RPF (*.RPF) These image formats are capable of containing any of the image channels mentioned above. All channels are contained within one file, including RGBA, as well as the auxiliary channels. These files are identified by the RLA or RPF file extension. Not all RLA or RPF files contain auxiliary channel information but most do. RPF files have the additional capability of storing multiple samples per pixel, so different layers of the image can be loaded for very complex depth composites. Fusion RAW (*.RAW) Fusion’s native RAW format is able to contain all of the auxiliary channels as well as other metadata used within Fusion. Creating Auxiliary Channels in Fusion The following nodes create auxiliary channels: Renderer 3D: Creates these channels in the same way as any other 3D application would, and you have the option of outputting every one of the auxiliary data channels that the Fusion page supports. Optical Flow: Generates Vector and Back Vector channels by analyzing pixels over consecutive frames to determine likely movements of features in the image. Disparity: Generates Disparity channels by comparing stereoscopic image pairs. Chapter – 12 Understanding Image Channels and Node Processing 2-75 Auxiliary Channels Explained Fusion is capable of using auxiliary channels, where available, to perform depth based compositing, to create masks and mattes based on Object or Material IDs, and for texture replacements. Tools that work with auxiliary channel information have been specifically developed to work with this data. Z-Depth Each pixel in a Z-Depth channel contains a value that represents the relative depth of that pixel in the scene. In the case of overlapping objects in a model, most 3D applications take the depth value from the object closest to the camera when two objects are present within the same pixel, since the closest object typically obscures the farther object. When present, Z-depth can be used to perform depth merging using the Merge node, or to control simulated depth-of-field blurring using the Depth Blur node. The rendered Z-Depth channel for the previous RGBA image Z-Coverage The Z-Coverage channel is used to indicate pixels in the Z-Depth that contains two objects. The value is used to indicate, as a percentage, how transparent that pixel is in the final depth composite. WARNING: Depth composites in Fusion that are based on images that lack a Z-Coverage channel, as well as a background RGBA channel, will not be properly anti-aliased Z-Coverage channel Chapter – 12 Understanding Image Channels and Node Processing 2-76 Background RGBA This channel contains the color values from the objects behind the pixels described in the Z-Coverage. Background RGBA Object ID Most 3D applications are capable of assigning ID values to objects in a scene. Each pixel in the Object ID channel will be identified by that ID number, allowing for the creation of masks. Object ID Material ID Most 3D applications are capable of assigning ID values to materials in a scene. Each pixel in the Material ID channel will be identified by that ID number, allowing for the creation of masks based on materials. Material ID Chapter – 12 Understanding Image Channels and Node Processing 2-77 UV Texture The UV Texture channels contain information about mapping coordinates for each pixel in the scene. This is used to apply textures wrapped to the object. UV Texture X, Y and Z Normals The X, Y and Z Normal channels contain information about each pixel’s orientation (the direction it faces) in 3D space. XYZ Normals XY Vector and XY BackVector The Vector channels indicates the pixel’s motion from frame to frame. It can be used to apply motion blur to an image as a post process or to track pixels over time for retiming. The XY Vector points to the next frame, while the XY BackVector points to the previous frame. XY Vector Chapter – 12 Understanding Image Channels and Node Processing 2-78 XYZ Position The XYZ Position channels indicate where each pixel is assigned; the XYZ position of its location is in 3D space, typically in world coordinates. This can be used, like Z-depth, for compositing in depth but can also be used for masking based on 3D position, regardless of camera transforms. For more information on using Position channels in Fusion, read Chapter 12, 3D. XYZ Position XY Disparity The XY Disparity channels indicate where each pixel’s corresponding matte can be found in a stereo image. Each eye, left and right, will use this vector to point to where that pixel would be in the other eye. This can be used for adjusting stereo effects, or to mask pixels in stereo space. XY Disparity Chapter – 12 Understanding Image Channels and Node Processing 2-79 Propagating Auxiliary Channels Ordinarily, auxiliary channels will be propagated along with RGBA image data, from node to node, among gray-colored nodes including those in the Blur, Filter, Effect, Transform, and Warp categories. Basically, most nodes that simply manipulate channel data will propagate (and potentially manipulate) auxiliary channels no problem. However, when you composite two image layers using the Merge node, auxiliary channels will only propagate through the image that’s connected to the background input. The rationale for this is that in most composites that include computer generated imagery, the background is most often the CG layer that contains auxiliary channels, while the foreground is a live-action greenscreen plate with subjects or elements that are combined against the background, which lack auxiliary channels. Viewing Auxiliary Channels You can view the Auxiliary Channels by selecting the desired channel from the Viewer’s toolbar or from the Viewer’s contextual menu. The Color Inspector SubV can also be used to read numerical values from all of the channels. Selecting a channel from the Viewer’s Toolbar Chapter – 12 Understanding Image Channels and Node Processing 2-80 Nodes That Use Auxiliary Channels The availability of Auxiliary channels opens up a world of advanced compositing functionality. This section describes every Fusion node that has been designed to work with images that contain Auxiliary channels. Merge In addition to regular compositing operations, Merge is capable of merging two or more images together using the Z-Depth, Z-Coverage, and BG RGBA buffer data. This is accomplished by enabling the Perform Depth Merge checkbox from the Channels tab. Depth Blur The Depth Blur tool is used to blur an image based on the information present in the Z-Depth. A focal point is selected from the Z-Depth values of the image and the extent of the focused region is selected using the Depth of Field control. Fog The Fog tool makes use of the Z-Depth to create a fog effect that is thin closer to the camera and thickens in regions farther away from the camera. You use the Pick tool to select the Depth values from the image and to define the Near and Far planes of the fog’s effect. Shader The Shader tool applies data from the RGBA, UV and the Normal channels to modify the lighting applied to objects in the image. Control is provided over specular highlights, ambient and diffuse lighting, and position of the light source. A second image can be applied as a reflection or refraction map. SSAO SSAO is short for Screen Space Ambient Occlusion. Ambient Occlusion is the lighting caused when a scene is surrounded by a uniform diffuse spherical light source. In the real world, light lands on surfaces from all directions, not from just a few directional lights. Ambient Occlusion captures this low frequency lighting, but it does not capture sharp shadows or specular lighting. For this reason, Ambient Occlusion is usually combined with Specular lighting to create a full lighting solution. The SSAO tool uses the Z-Depth channel, but requires a Camera3D input. Texture The Texture tool uses the UV channels to apply an image from the second input as a texture. This can replace textures on a specific object when used in conjunction with the Object ID or Material ID masks. Shadow The Shadow tool can use the Z-Depth channel for a Z-Map. This allows the shadow to fall onto the shape of the objects in the image. Chapter – 12 Understanding Image Channels and Node Processing 2-81 Vector Motion Blur Using the forward XY Vector channels, the Vector Motion Blur tool can apply blur in the direction of the velocity, creating a motion blur effect. Vector Distortion The forward XY Vector channels can be used to warp an image with this tool. Time Speed and Time Stretcher These tools can use the Vector and BackVector channels to retime footage. New Eye For stereoscopic footage, New Eye uses the Disparity channels to create new viewpoints or to transfer RGBA data from one eye to the other. Stereo Align For stereoscopic footage, the Disparity channels can be used by Stereo Align to warp one or both of the eyes to correct misalignment or to change the convergence plane. Smooth Motion Smooth Motion uses Vector and Back Vector channels to blend other channels temporally. This can remove high frequency jitter from problematic channels such as Disparity. Volume Fog Volume Fog is a raymarcher that uses the Position channels to determine ray termination and volume dataset placement. It can also use cameras and lights from a 3D scene to set the correct ray start point and Illumination parameters. Volume Mask Volume Mask uses the Position channels to set a mask in 3D space as opposed to screen space. This allows a mask to maintain perfect tracking through a camera move. Custom Tool, Custom Vertex 3D, pCustom The “Custom” tools can sample data from the auxiliary channels per pixel, vertex, or particle and use that for whatever processing you would like. Lumakeyer The Lumakeyer tool can be used to perform a key on the Z-Depth channel by selecting the Z-Depth in the channel drop down list. Disparity to Z, Z to Disparity, Z to WorldPos These tools use the inherent relationships between depth, position, and disparity to convert from one channel to another. Chapter – 12 Understanding Image Channels and Node Processing 2-82 Copy Aux The Copy Aux tool can copy auxiliary channels to RGB and then copy them back. It includes some useful options for remapping values and color depths, as well as removing auxiliary channels. Channel Boolean The Channel Boolean tool can be used to combine or copy the values from one channel to another in a variety of ways. TIP: The Object ID and Material ID auxiliary channels can be used by some tools in Fusion to generate a mask. The “Use Object” and “Use Material” settings used to accomplish this are found in the Settings tab of that node’s controls in the Inspector. Chapter – 12 Understanding Image Channels and Node Processing 2-83 Chapter 13 Learning to Work in the Fusion Page This chapter is a grand tour of the basics of the Fusion page, walking you through the process of shepherding a clip from the Edit page to the Fusion page, and then working in the Node Editor to create some simple effects. Subsequent topics build upon these basics to show you how to use the different features in Fusion to accomplish common compositing and effects tasks. In the process you’ll learn how node trees are best constructed, and how to use the different panels of the Fusion page together to work efficiently. Chapter – 13 Learning to Work in the Fusion Page 2-84 Contents What’s a Composition? 2-87 Moving From the Edit to the Fusion Page 2-87 How Nodes Are Named 2-88 Applying and Masking Effects 2-88 Adding a Node to the Tree 2-89 Editing Parameters in the Inspector 2-89 Replacing Nodes 2-91 Adjusting Fusion Page Sliders 2-91 Masking Node Effects 2-92 Color Management in the Public Beta 2-95 Compositing Two Clips Together 2-97 Adding Additional Media to Compositions 2-97 Automatically Creating Merge Nodes 2-98 Adjusting the Timing of Clips Added From the Media Pool 2-99 Fixing Problem Edges in a Composite 2-100 Composite Modes and the Corner Positioner 2-101 Setting Up the Initial Composite 2-101 Controlling Which Node You See in the Viewer 2-103 Adding the Corner Positioner Node With a Search 2-104 Warping the Image With the Corner Positioner Node 2-105 Toggling On-screen Control Visibility 2-106 Navigating the Viewer 2-107 Using the Screen Composite Mode in the Merge Node 2-107 Tweaking Color in the Foreground Layer 2-108 Creating and Using Text 2-110 Creating Text Using the Text+ Node 2-110 Styling and Adjusting Text 2-111 Using One Image’s Alpha Channel in Another Image 2-112 Using Transform Controls in the Merge Node 2-116 Match Moving Text With Motion Tracking 2-117 Adding a Layer We Want to Match Move 2-117 Setting Up to Track 2-118 A Simple Tracking Workflow 2-119 Connecting Motion Track Data to Match Move 2-122 Offsetting the Position of a Match Moved Image 2-123 Chapter – 13 Learning to Work in the Fusion Page 2-85 Using Paint and Planar Tracking 2-125 Using a Planar Tracker to Steady a Subject to Paint 2-125 Painting Over Blemishes 2-128 Inverting the Steady Effect to Put the Motion Back In 2-131 Fixing the Edges by Only Using the Fixed Part of the Frame 2-133 Building a Simple Green Screen Composite 2-137 Organizing Clips in the Edit Page to Create a Fusion Clip 2-137 Pulling a Greenscreen Key Using the Delta Keyer 2-138 Using the Transform Node to Resize a Background 2-142 Masking a Graphic 2-143 Animating an Image Using Keyframes 2-146 Animating a Parameter in the Inspector 2-146 Using the Spline Editor 2-147 Congratulations 2-149 Chapter – 13 Learning to Work in the Fusion Page 2-86 What’s a Composition? A “composition” describes the collection of nodes that creates an effect in the Fusion page, just as a “grade” describes the collection of nodes that creates a color adjustment or look on the Color page. The relationship between the Edit page and the Fusion page, at a basic level, is similar to the relationship between the Edit page and the Color page. Every clip can have a grade applied to it in the Color page, and similarly every clip can have a composition applied to it in the Fusion page. If you use the Fusion page to add effects or do any compositing at all, a badge appears on that clip in all timelines to show that clip has a composition applied to it. Clips with Fusion page compositions have a Fusion badge to the right of the name Moving From the Edit to the Fusion Page Whenever you want to create a composite using a clip from the Edit page, the simplest way to work is to just move the playhead so it intersects the desired clip in the Edit page timeline, make sure the clip you want to composite is the topmost clip of any superimposed stack of clips (whichever clip you see in the Timeline Viewer is the one you’ll be compositing with), and then open the Fusion page. Positioning the playhead over a clip you want to use in a composition In the Fusion page, you should see a single selected MediaIn1 node representing only the top-most clip you where parked on in the Edit page, and that image should be showing in the Viewer thanks to the MediaOut1 node automatically being loaded in the Viewer (the Viewer buttons visible underneath that node confirm this). Any clips that were underneath that clip are ignored when you work this way, because the idea is you’re only doing a quick fix to the current clip at the position of the playhead. Chapter – 13 Learning to Work in the Fusion Page 2-87 How the Fusion page appears when you first open it while the playhead is on a new clip The playhead should still be on the same frame you were parked on in the Edit page, except now it’s on the equivalent source frame in the Time Ruler underneath the Viewer that represents that clip’s media. Yellow markers indicate the range of the current clip that appears in the Timeline, while the source clip’s handles extend to the left and right. Lastly, the selected MediaIn1 node displays its parameters in the Inspector to the right. In the Node Editor at the bottom, the MediaIn1 node is connected to a MediaOut1 node. If this is all you see, there is no effect yet applied to this clip. It’s only when you start adding nodes between MediaIn and MediaOut that you begin to assemble a composition. At this point, you’re ready to start compositing. How Nodes Are Named While the documentation refers to nodes by their regular name, such as “MediaIn,” the actual names of nodes in the Fusion Node Editor have a number appended to them, to indicate which node is which when you have multiple instances of a particular type of node. Applying and Masking Effects Let’s begin by looking as some very simple effects, and build up from there. Opening the Effects Library, then clicking the Disclosure control to the left of Tools, reveals a list of categories containing all the effects nodes that are available in Fusion. As mentioned before, each node does one thing, and by using these nodes in concert you can create extremely complex results from humble beginnings. Clicking the Effect category reveals its contents. For now, we’re interested in the TV effect. Chapter – 13 Learning to Work in the Fusion Page 2-88 Browsing the Effect category to find the TV node Adding a Node to the Tree Assuming the MediaIn node is selected in the Node Editor, clicking once on the TV node in the Effects Library automatically adds that node to the node tree to the right of the selected node, and it immediately takes effect in the Viewer thanks to the fact that the MediaOut1 node is what’s loaded in the Viewer, since that means that all nodes upstream of the MediaOut1 node will be processed and shown. A new node added from the Effects Library There are many other ways of adding nodes to your node tree, but it’s good to know how to browse the Effects Library as you get started. Editing Parameters in the Inspector Looking at the TV effect in the Viewer, you may notice a lot of transparency in the image because of the checkerboard pattern. If you don’t see the checkerboard pattern in the Viewer, it might be turned off. You can turn it on by clicking the Viewer option menu and choosing Checker Underlay. To improve the effect, we’ll make an adjustment to the TV1 node’s parameters in the Inspector at the left. Whichever node is selected shows its controls in the Inspector, and most nodes have several panels of controls in the Inspector, seen as little icons just underneath that node’s title bar. Chapter – 13 Learning to Work in the Fusion Page 2-89 The Inspector showing the parameters of the TV effect Clicking the last panel opens the Settings panel. Every node has a Settings panel, and this is where the parameters that every node shares, such as the Blend slider and RGBA checkboxes, are found. These let you choose which image channels are affected, and let you blend The Settings panel, which has channel limiting and mask handling controls that every node shares In our case, the TV effect has a lot of transparency because the scan lines being added are also being added to the alpha channel, creating alternating lines of transparency. Turning the Alpha checkbox off results in a more solid image, while opening the Controls panel (the first panel) and dragging the Scan Lines slider to the right to raise its value to 4 creates a more visible television effect. (Left) The original TV effect, (Right) Modifications to the TV effect to make the clip more solid Chapter – 13 Learning to Work in the Fusion Page 2-90 Replacing Nodes That was fun, but having previewed this effect, we decide we want to try something different with this shot. Going back to the Effect category of the Effects Library, there is a Highlight node we can use to add some pizazz to this shot, instead of the TV node. Instead of clicking the Highlight node, which would add it after the currently selected node, we’ll drag and drop it on top of the TV1 node in the Node Editor. A dialog appears asking “Are you sure you want to replace TV1 with Highlight?” and clicking OK makes the replacement. Dragging a node from the Effects Library onto a node in the Node Editor to replace it A Highlight1 node takes the TV node’s place in the node tree, and the new effect can be seen in the Viewer, which in this image’s case consists of star highlights over the lights in the image. Incidentally, another way you can replace an existing node with another type of node in the Node Editor is to right-click a node you want to replace, and choose the new node you want from the Replace Tool submenu of the contextual menu that appears. Right-clicking a node to use the contextual menu Replace Node submenu It’s time to use the Inspector controls to customize this effect, but first, let’s take a look at how sliders in the Fusion page differ somewhat from sliders on other pages in DaVinci Resolve. Adjusting Fusion Page Sliders When you drag a slider in the Fusion page Inspector, in this case the “Number of Points” slider, a little dot appears underneath it. This dot indicates the position of the default value for that slider, and also serves as a reset button if you click it. Chapter – 13 Learning to Work in the Fusion Page 2-91 Adjusting a slider reveals a reset button underneath it Each slider is limited to a different range of minimum and maximum values that is particular to the parameter you’re adjusting. In this case, the “Number of Points” slider maxes out at 24. However, you can re-map the range of many (not all) sliders by entering a larger value in the number field to the right of that slider. Doing so immediately repositions the slider’s controls to the left as the slider’s range increases to accommodate the value you just entered. Entering a larger value to expand the range over which a slider will operate Masking Node Effects Going back to the Length slider, increasing its value gives us a nice big flare. The Highlight effect with a raised Length value (zoomed in) This is a nice effect, but maybe we only want to apply it to the car in the foreground, rather than to every single light in the scene. This can be accomplished using a Mask node connected to the Effect input of the Highlight node. The Effect Mask input is a blue input that serves a similar function to the KEY input of nodes in the Color page; it lets you use a mask or matte to limit that node’s effect on the image, like a secondary adjustment in color correction. Most nodes have an Effects Mask input, and it’s an enormously useful technique. However, there’s another node input that’s more interesting, and that’s the gray Highlight Mask input on the bottom of the node. This is an input that’s specific to the Highlight node, and it lets you use a mask to limit the part of the image that’s used to generate the Highlight effect. The blue Effect input of a node is on top, and the gray Highlight Mask input that’s specific to the Highlight node is on the bottom Chapter – 13 Learning to Work in the Fusion Page 2-92 Adding a Mask Node To see the results of using either of these two inputs, let’s add a mask, this time using the toolbar, which presents a collection of frequently-used mask nodes that we can quickly create. Clicking the Ellipse button on the Toolbar With the Highlight node selected already, clicking the Ellipse button (the circle) automatically creates an Ellipse1 node that’s connected to the blue Effect Mask input. Creating new masks while a node is selected always auto-connects to that node’s Effect Mask input as the default behavior. Automatically connecting an Ellipse node to the blue Effect Mask Adjusting Mask Nodes Masks, in the Fusion page, are shapes you can either draw or adjust that have a special single-channel output that’s meant to be connected to specialized mask inputs, to either create transparency or limit effects in different ways as described above. With the Ellipse1 node connected and selected, a round on-screen control appears in the Viewer that can be adjusted in different ways. Drag on the edges of the mask to reshape it Drag the center handle to reposition it freely Drag the up or right arrows to reposition it constrained vertically or horizontally Drag the top, bottom, left, or right sides of the ellipse to stretch it vertically or horizontally Drag any of the corners of the ellipse to resize it proportionally. Resizing the ellipse to hug only the headlights of the main car, you can see that using the Effect Mask cuts off the long flares we’ve created, because this masks the final effect to reveal the original image that’s input into that node. Chapter – 13 Learning to Work in the Fusion Page 2-93 The result of using the Effect Mask input Reconnecting Node Connections to Different Inputs for a Different Result This isn’t satisfactory, so we drag the connection line attaching the Ellipse node off the Effect Mask input and onto the Highlight Mask input underneath. It’s easy to reconnect previously connected nodes in different ways simply by dragging the second half of any connection (it highlights when you hover the pointer over it) to any other node input you want to connect to. Dragging a connection from one node input to another After you make the connection, the connection line goes back to the top of the node, and the top connection is now gray. This is because node inputs in the Fusion page automatically rearrange themselves to keep the node tree tidy, preventing connection lines from overlapping nodes unnecessarily and creating a mess. This may take a bit of getting used to, but once you do, you’ll find it an indispensable behavior. The Ellipse node now connected to the Highlight Mask input, which has moved to the top of the node to keep things neat Chapter – 13 Learning to Work in the Fusion Page 2-94 Now that the Ellipse1 node is connected to the Highlight Mask, the tight mask we’ve created just around the car headlights restrict the node in a different way. The Highlight Mask lets you restrict which part of the image is used to trigger the effect, so that only the masked car headlights will generate the Highlight effect in this filter. The result is that the flares of the Highlight effect themselves are unhindered, and stretch well beyond the boundaries of the mask we’ve created. The highlight effect is uncropped because the effect is being limited via the Highlight Mask input, rather than the Effect Mask input Unlike nodes on the Color page that have largely the same inputs and outputs, nodes on the Fusion page may have any number of inputs that are particular to what that node does. This example should underscore the value of getting to know each node’s unique set of inputs, in order to best control that node’s effect in the most appropriate way. Color Management in the Public Beta If you want to follow a professional compositing workflow of converting media gamma to linear as it comes into the Fusion page, and then back to the original timeline gamma prior to the MediaOut node, at the time of this writing, you must do this manually using either CineonLog or FileLUT nodes. The CineonLog node is good if you’re converting from or to a log-encoded gamma. The FileLUT node is good if you’re converting from or to a gamma of 2.4, since it can use the LUTs that are available in the /Library/Application Support/Blackmagic Design/ DaVinci Resolve/LUT/VFX IO/ directory (on macOS). What this looks like in the node tree is that each MediaIn node will have a CineonLog or FileLUT node attached to it doing a conversion to linear gamma, while the MediaOut node will have a CineonLog or FileLUT node attached just before it doing a conversion from linear gamma to the timeline gamma, whatever that happens to be. All other operations in the composition must be applied between these two conversions. Converting all images to Linear coming into the composite, and converting out of linear out of the composite, in this case using CineonLog nodes at the beginning and end Chapter – 13 Learning to Work in the Fusion Page 2-95 TIP: The CineonLog node can be found in the Film category of the Effects Library, while the FileLUT node is found in the LUT category. In this example, the timeline gamma is BMD Film 4.6K, which is a log-encoded gamma, so the first CineonLog node does a Log to Lin conversion from this format. The second CineonLong node then does a Lin to Log conversion to this format, in order to move the image data out of the Fusion page in the way DaVinci Resolve expects. (Left) The first CineonLog node does a Log to Lin conversion, (Right) the second CineonLog node does a Lin to Log conversion While this is happening, you’ll want to set your Viewer to a gamma setting that lets you see the image as it will look when the audience sees it (more or less). You can do this using the Viewer LUT button. Click to turn it on, and then choose a setting from the VFX IO submenu that you want to preview your composite with as you work, such as Linear to 2.4, corresponding to the BT.1886 standard for gamma used for SDR HD output. The Viewer LUT button (enabled) and pop-up menu Chapter – 13 Learning to Work in the Fusion Page 2-96 This is a bit of set-up, it’s true, but it will provide superior compositing results, especially for compositions with filtering and lighting effects. NOTE: Eventually, the MediaIn and MediaOut nodes will include built-in functionality for doing gamma and color space conversions, similarly to the standalone version of Fusion. Furthermore, the Fusion page will at some point in the future be governed by Resolve Color Management (RCM), similarly to the Edit and Color pages, so many of the examples in this chapter will omit these nodes from the node trees. Compositing Two Clips Together As entertaining as it is adding individual nodes to create simple effects, eventually you need to start adding additional layers of media in order to merge them together as composites. Let’s turn our attention to another composition in which we need to combine a background clip with a foreground clip that’s been handed to us that already has a built-in alpha channel, to see a simple composite in action. Adding Additional Media to Compositions You’ll often find that even though you start out wanting to do something relatively simple, you end up adding additional media to create the effect that you need. For this reason, you can open the Media Pool on the Fusion page and drag clips directly to the Node Editor to add them to your node tree. Clicking the Media Pool button in the UI Toolbar opens the Media Pool, which if you’re already familiar with DaVinci Resolve, is the same Media Pool that’s now found in every page except for the Deliver page. The Media Pool shares the same area with the Effects Library, so if you have them both open at the same time, they’ll be stacked one on top of another The Media Pool as seen in the Fusion Page Chapter – 13 Learning to Work in the Fusion Page 2-97 If you drag a clip from the Media Pool to an empty area of the Node Editor, you’ll add an unconnected MediaIn2 node (incremented to keep it unique) that you can then connect in any way you want. Automatically Creating Merge Nodes However, there’s a shortcut if you want to connect the incoming clip immediately to your node tree as the top layer of a composite, and that’s to drag the clip right on top of any connection line. When you drop the resulting node, this automatically creates a Merge1 node, the “background input” of which is connected to the next node to the left of the connection you dropped the clip onto, and the “foreground input” of which is connected to the new MediaIn2 node that represents the clip you’ve just added. (Left) Dragging a node from the Media Pool onto a connection, (Right) Dropping it to create a Merge node composite The Fusion page Node Editor is filled with shortcuts like this to help you build your compositions more quickly. Here’s one for if you have a disconnected node that you want to composite against another node with a Merge node. Drag a connection from the output of the node you want to be the foreground layer, and drop it on top of the output of the node you want to be the background layer, and a Merge node will be automatically created to build that composite. Remember, background inputs are orange, and foreground inputs are green. (Left) Dragging a connection from a disconnected node to another node’s output (Right) Dropping it to create a Merge node composite Adding Clips to Fusion From the File System If you drag clips from the file system directly into the Node Editor, they’ll be added to the Media Pool automatically. So, if you have a library of stock animated background textures and you’ve just found one you want to use using your file system’s search tools, you can simply drag it straight into the Node Editor and it’ll be added to the currently selected bin of the Media Pool. Chapter – 13 Learning to Work in the Fusion Page 2-98 Adjusting the Timing of Clips Added From the Media Pool Because the MediaIn2 node that’s connected to the Merge1 node’s foreground input has an alpha channel, this simple Merge node composite automatically creates a result that we can see in the Viewer, but the way the two clips line up at the beginning of the composition’s range in the Time Ruler is not great, because the MediaIn2 node’s clip is being lined up with the very first frame of the MediaIn1 clip’s handles, rather than the first frame of the actual composition range as seen by the yellow marks in the Time Ruler. The composite is good, but the timing of the foreground clip relative to the background clips is not ideal This a not uncommon issue with clips you add from the Media Pool in the Fusion page, because those clips were never edited into the Timeline in the Edit page where they could be properly timed and trimmed relative to the other clips in the composition. Fortunately, you can slip clips and resize their In and Out points using the Keyframes Editor, which can be opened via a button in the UI Toolbar. The Keyframes Editor shows each MediaIn and Effect node as a bar in a vertical stack that shows you the relative timing of each clip and effect. Keep in mind that the vertical order of these layers is not indicative of which layers are in front of others, as that is defined by layer input connections in the Node Editor. The layers displayed in the Keyframe Editor are only intended to show you the timing of each composited clip of media. In this case, we can see that the MediaIn2 node is offset to the left, so it’s easy for us to drag it to the right, watching the image in the Viewer, until the frame at the composition in point is what we want. (Top) The original stack of layers, (Bottom) Sliding the MediaIn2 layer to line it up better with the MediaIn1 layer Chapter – 13 Learning to Work in the Fusion Page 2-99 As a result, the MediaIn2 clip lines up much better with the MediaIn1 clip. After sliding the MediaIn2 clip to improve its alignment with the other clip in the composite Fixing Problem Edges in a Composite Most of the time, the Merge node does a perfectly good job when handed a foreground image with premultiplied alpha transparency to composite against a solid background image. However, from time to time, you may notice a small bit of fringing at the edge of the border of a foreground element and transparent area, such as seen in the following close-up. This slight lightening at the edge is a tell-tale sign that the clip probably wasn’t pre-multiplied. But this is something that’s easily fixed. A bit of fringing at the edge of a foreground element surrounded by transparency Click to select the Merge node for that particular composite, and look for the Subtractive/ Additive slider. The Subtractive/Additive slider, which can be used to fix or improve fringing in composites Drag the slider all the way to the left, to the Subtractive position, and the fringing disappears. (Left) A clip with alpha exhibits fringing, (Right) Fixing fringing by dragging the Subtractive/Additive slider to the left Chapter – 13 Learning to Work in the Fusion Page 2-100 The Subtractive/Additive slider, which is only available when the Apply Mode is set to Normal, controls whether the Normal mode performs an Additive merge, a Subtractive merge, or a blend of both. This slider defaults to Additive merging, which assumes that all input images with alpha transparency are pre-multiplied (which is usually the case). If you don’t understand the difference between Additive and Subtractive merging, here’s a quick explanation: An Additive merge, with the slider all the way to the right, is necessary when the foreground image is pre-multiplied, meaning that the pixels in the color channels have been multiplied by the pixels in the alpha channel. The result is that transparent pixels are always black, since any number multiplied by 0 is always going to be 0. This obscures the background (by multiplying with the inverse of the foreground alpha), then simply adds the pixels from the foreground. A Subtractive merge, with the slider all the way to the left, is necessary if the foreground image is not pre-multiplied. The compositing method is similar to an Additive merge, but the foreground image is first multiplied by its own alpha, to eliminate any background pixels outside the alpha area. The Additive/Subtractive slider lets you blend between two versions of the merge operation, one Additive and the other Subtractive, to find the best combination for the needs of your particular composite. Blending between the two is an operation that is occasionally useful for dealing with problem composites that have edges that are calling attention to themselves as either too bright or too dark. For example, using Subtractive merging on a pre-multiplied image may result in darker edges, whereas using Additive merging with a non-premultiplied image will cause any non-black area outside the foreground’s alpha to be added to the result, thereby lightening the edges. By blending between Additive and Subtractive, you can tweak the edge brightness to be just right for your situation. Composite Modes and the Corner Positioner In this next compositing example, we’ll explore how you can use the Corner Positioner node to cornerpin warp a composited layer into place as a screen replacement. Then we’ll use a composite mode in the Blend node to refine the screen replacement effect to incorporate real reflections from the scene. Setting Up the Initial Composite The base image in the MediaIn1 node is a clip that’s been zoomed into in the Edit page. When you use the Transform, Cropping, or Lens Correction controls for a clip in the Edit page Inspector, those adjustments are passed along as the initial state of the image in the Fusion page, allowing for some prep work to be done in the Edit page, if necessary. Chapter – 13 Learning to Work in the Fusion Page 2-101 Adjusting the Edit sizing of a clip before moving it into the Fusion page for compositing Because this particular example uses the Screen Composite mode to do a composite, we’ll start by setting up some routine color management in the node tree, to illustrate how this should be handled. Taking a Shortcut by Copying and Pasting Nodes In the Fusion page, this first clip has been converted to and from linear gamma using a FileLUT node set to use the “Gamma 2.4 to Linear.cube” from the /Library/Application Support/ Blackmagic Design/DaVinci Resolve/LUT/VFX IO directory. However, after dragging and dropping the video image we need to insert into the screen onto a connection to automatically add it connected to a Merge1 node, we find we need to add another copy of the same FileLUT node after the new MediaIn2 node. Happily, this is easy to do by selecting and copying the FileLUT1 node that’s connected to the MediaIn1 node (Command-C), then selecting the MediaIn2 node, and pasting (Command-V). When you paste one or more nodes while a node is selected in the Node Editor, the nodes you paste are inserted onto the connection line from the selected node’s output. You can tell when a node has been copied and pasted because it shares the same name as the copied name, but with a “_#” appended to it. Copying a node from one part of the node tree and pasting to insert it into after a selected node Chapter – 13 Learning to Work in the Fusion Page 2-102 If we then select the Merge1 node, we can paste another instance of this FileLUT node to come just before the MediaOut1 node, setting its LUT File to the “Linear to Gamma 2.4.cube” LUT that’s also found in the /Library/Application Support/Blackmagic Design/DaVinci Resolve/LUT/ VFX IO directory. TIP: If you paste one or more nodes while no nodes are selected, then you end up pasting nodes that are disconnected. However, to control where disconnected nodes will appear when pasted, you can click the place in the Node Editor where you’d like pasted nodes to appear, and when you paste, the nodes will appear there. Controlling Which Node You See in the Viewer Since we’re doing gamma conversions to media coming into and going out of the Fusion page, it’s no longer suitable to View the MediaOut node as we work, because the Viewer is currently only set up to convert the linear image data that’s in-between the two sets of FileLUT nodes to something normal for you to look at (such as gamma 2.4). Happily, there are a wide variety of ways you can load a particular node into the Viewer to see what you’re doing as you work: Hover the pointer over a node, and click one of two buttons that appear at the bottomleft of the node. Click once to select a node, and press 1 (for the left viewer) or 2 (for the right viewer). Right-click a node and choose View On > None/LeftView/RightView in the contextual menu. Drag a node and drop it over the viewer you’d like to load it into (this is great for tablet users). Using any of these methods, we load Merge1 into the Viewer. Loading a node from the middle of the node tree into the Viewer to see a specific node you’re working on Chapter – 13 Learning to Work in the Fusion Page 2-103 We can tell which node is loaded into the Viewer because of the Viewer indicators/buttons at the bottom left of the node. Not only is this a visual indication of which node is being viewed, but these buttons can be clicked to load that node into the left or right Viewer, if you go into Dual-viewer mode. A pair of buttons at the bottom-left of nodes that are loaded into the Viewer let you see which node is being viewed, as well as giving you a click-target for reassigning that node to another Viewer Adding the Corner Positioner Node With a Search Now that we have a foreground image composited over the background image of the computer screen, it’s time to reposition the foreground layer to fit into the screen. To do so, we’ll use the Corner Positioner node, from the Warp category, which is the main node for doing cornerpinning. To add this to the node tree, we’ll use a different method to search for the node we want right from the Node Editor. First, select the node you want to insert a new node after. In this case, we want to cornerpin the image from the MediaIn2 node, so we’ll select the FileLUT_1 node that’s attached to it. Selecting a node you want to add another node behind Next, pressing Shift-Spacebar opens the Select Tool dialog. Once it appears, just start typing the first few letters of the name of the node you’re looking for to find it in a master list of every node in the entire Fusion page. In this case, you’re looking for the CornerPositioner node, so type “cor” and the list of nodes will shrink to two, with the one we’re looking for being selected. Chapter – 13 Learning to Work in the Fusion Page 2-104 Pressing Shift-Spacebar opens the Select Tool dialog for quickly finding and adding new nodes With the node we’re looking for found and selected in the Select Tool dialog, pressing the Return key inserts the new Corner Positioner node after the previously selected node and closes the Select Tool dialog. The CornerPositioner node added to cornerpin the foreground image prior to the Merge operation Warping the Image With the Corner Positioner Node The Corner Positioner node is a node in the Warp category of the Effects Library that lets you do absolute positioning at four corner points to fit an image within a rectangular region to into a scene. Immediately upon adding this node, a default cornerpin operation is applied to the image to show that it’s being manipulated. Chapter – 13 Learning to Work in the Fusion Page 2-105 The Corner Positioner node adds a default transform to the image Using the on-screen control points, we can now warp the image by dragging each corner to fit within the computer screen. Using the CornerPositioner node to fit the video image to the screen it’s replacing Toggling On-screen Control Visibility It’s worth knowing that you can toggle the visibility of on-screen controls using Show Controls in the Viewer Option Menu. You might find it useful to hide on-screen controls if they’re getting in the way of seeing the image you’re adjusting, but if you’ve added an effect and you don’t see any controls available for adjusting it, you’ll know you need to turn this option on. Show Controls in the Option menu toggles on-screen control visibility on and off Chapter – 13 Learning to Work in the Fusion Page 2-106 Navigating the Viewer As you work, you may find that parts of the image you want to work on extend off screen. To deal with this, there are a few ways of panning and zooming around the Viewer. Middle click and drag to pan around the Node Editor. Press Command and Shift and drag to pan around the Node Editor. Hold the Middle and Left buttons down simultaneously and drag to zoom into or out of the Node Editor. Hold the Command key down and use your scroll wheel to zoom in and out of the Node Editor. Using the Screen Composite Mode in the Merge Node Once the foreground input image is fit to the screen, we have an opportunity to create a more convincing composite by taking advantage of the reflections of the scene on the front of the screen, and using the screen composite mode to make the foreground image look more like a reflection. The Merge node has a variety of controls built into it for creating just about every compositing effect you need, including an Apply Mode pop-up menu that has a selection of composite modes you can use to combine the foreground and background layers together, and a Blend slider you can use to adjust how much of the foreground input image to merge with the background. Adjusting the Apply Mode and Blend slider of the Merge node in the Inspector NOTE: The Subtractive/Additive slider disappears when you choose any other Apply Mode option besides Normal, because the math would be invalid. This isn’t unusual; there are a variety of controls in the Inspector which hide themselves when not needed or when a particular input isn’t connected. The screen node is perfect for simulating reflections, and lowering Blend a bit lets you balance the coffee cup reflections from the display in the background with the image in the foreground. It’s subtle, but helps sell the shot. Chapter – 13 Learning to Work in the Fusion Page 2-107 (Left) The original composite, (Right) The composite using the Screen apply mode TIP: You may have noticed that the Merge node also has a set of Flip, Center, Size, and Angle controls that you can use to transform the foreground image without needing to add a dedicated Transform node. It’s a nice shortcut for simplifying node trees large and small. Tweaking Color in the Foreground Layer It’s as important to make sure that the color matches between two images being composited together as it is to create a convincing blend, and for this reason the Fusion page has a whole category of color adjustment tools available in the Color category of the Effects Library. In fact, the ColorCorrector, ColorCurves, HueCurves, and Brightness/Contrast nodes are considered so important they appear in the Toolbar. Frequently used Color nodes in the Toolbar In this particular image, the color of the foreground image on the computer screen is just a bit green and oversaturated, and the view out the window is a bit magenta. However, these problems are easily overcome using a HueCurves node from the Toolbar. Selecting the CornerPositioner node we added, clicking the HueCurves button on the Toolbar adds that node between the CornerPositioner and the Merge node. Adding the HueCurves node to make a correction to the foreground image Chapter – 13 Learning to Work in the Fusion Page 2-108 The HueCurves node exposes a curve control in the Inspector with options for adjusting 9 kinds of curves, each overlapping the others for simultaneous adjustment. Turning on first the Hue checkbox to make adjustments, and then the Sat checkbox in the Inspector, these two curves can be simultaneously adjusted to push the green towards a healthier red in the skin tones of both the man and the woman, to desaturate the red, yellow, and green a bit, and to push the magenta outside the window to more of a warm orange light, to make the foreground seem like a natural part of the image. The controls of the HueCurves node, adjusted to correct the screen replacement image The result is subtle, but it’s a much more convincing composite. (Left) The uncorrected foreground, (Right) Using a hue curve node to adjust the image for a better composite Chapter – 13 Learning to Work in the Fusion Page 2-109 Creating and Using Text In this next example, we’ll take a look at how to create a simple text object using the Text+ node. Then, we’ll see how to use the text generator’s alpha channel in another image to create a more complex composite. Creating Text Using the Text+ Node The Text+ node is the primary tool for creating 2D text in the Fusion page. This is also the new text generator that has become available in the Edit page, and because it’s so ubiquitous, it appears in the Toolbar. The Text+ node is an incredibly deep tool for creating text effects, with six panels of controls for adjusting everything from text styling, to different methods of layout, to a variety of shading controls including fills, outlines, shadows, and borders. As sophisticated a tool as this is, we’ll only be scratching the surface in this next demonstration. With the MediaIn1 node that will serve as our background selected in the Node Editor, clicking the Text+ button automatically creates a new Text+ node connected to the foreground input of a Merge node (Top) Selecting a node you want to append another node to, (Bottom) Clicking the Text+ button on the Toolbar automatically creates a Merge composite with the text as the foreground input connection Selecting the Text1 node opens the default “Text” panel parameters in the Inspector, and it also adds a toolbar at the top of the Viewer with tools that are specific to that node. Clicking on the first tool at the left lets us type directly into the Viewer, so we type “SCHOOLED” into the Styled Text field, since that’s the title of the program we’re working on (in case you didn’t know). The Viewer toolbar for the Text node with tools for text entry, kerning, and outline controls Chapter – 13 Learning to Work in the Fusion Page 2-110 The text appears in the Viewer, superimposed against the background clip. Onscreen controls appear that let us rotate (the circle) and reposition (the red center handle and two arrows) the text, and we can see a faint cursor that lets us edit and kern the text using other tools in the Viewer toolbar. At this point, we’ve got our basic title text. Text that’s been typed into the Viewer, with on-screen text transform controls Styling and Adjusting Text Now we need to style the text to suit our purposes, so we’ll use the controls in the Inspector, increasing Size and decreasing Tracking to move the letters closer together so they can be made larger. The restyled text TIP: Holding the Command key down while dragging any control in the Inspector “gears down” the adjustment so that you can make smaller and more gradual adjustments. The result has somewhat uneven kerning, but we can adjust that. Selecting the manual kerning tool in the Viewer toolbar (second tool from the left) reveals small red dots underneath each letter of text. The Manual Kerning tool in the Viewer toolbar Chapter – 13 Learning to Work in the Fusion Page 2-111 Clicking a red dot under a particular letter puts a kerning highlight over that letter. Here are the different methods you can use to make Manual kerning adjustments: Option-drag the red dot under any letter of text to adjust that character’s kerning while constraining letter movement to the left and right. You can also drag letters up and down for other effects. Depending on your system, the kerning of the letter you’re adjusting might not update until you drop the red dot in place. If you don’t like what you’ve done, you can open the Advanced Controls in the Inspector, and clear either the kerning of selected letters, or all manual kerning, before starting over again. Option-dragging the little red dot revealed by the Manual Kerning tool to manually adjust kerning left or right So there we go, we’ve now got a nice title, styled using the Viewer tools and Inspector controls on the Text panel. This looks good, but we’ve got much grander designs. Using One Image’s Alpha Channel in Another Image We’re not going to use the text we’ve created as a title directly. Instead, we’re going to use the text as a matte to cut these letters out of another layer we’ll be using for texture. So, first we’ll drag another clip, of a chalkboard covered with math, from the Media Pool to the Node Editor as a disconnected MediaIn2 node. Disconnecting and Reconnecting Nodes Now we need to do a little rearranging. Moving the Merge1 node up, then clicking the last half of the connection from the Text1 node to the Merge foreground input to disconnect it. Chapter – 13 Learning to Work in the Fusion Page 2-112 (Top) Clicking the second half of a connection to disconnect it (Bottom) After Next, we’ll drag a connection from the MediaIn2 node onto the Merge1 node’s foreground input, so the entire Viewer becomes filled with the chalkboard (assuming we’re still viewing the MediaOut node). At this point, we need to insert the Text1 node’s image as an alpha channel into the MediaIn2 node’s connection, and we can do that using a MatteControl node. The updated composite, with two video images connected and the text node disconnected Using Matte Control Nodes Selecting the MediaIn2 node, we click the Matte Control button of the Toolbar to add it between the MediaIn2 and Merge1 nodes (to tidy things up, I’ve moved the nodes around a bit in the screenshot). The MatteControl node has many, many uses. Among them is taking one or more masks, mattes, or images that are connected to the garbage matte, solid matte, and/or foreground inputs, combining them, and using the result as an alpha channel for the image that’s connected to the background input. It’s critical to make sure that the image you want to add an alpha channel to is connected to the background input of the MatteControl node, as seen in the following screenshot, or the MatteControl node won’t work. Chapter – 13 Learning to Work in the Fusion Page 2-113 The second image properly connected to the Matte Control node’s background input With this done, we’ll connect the Text1 node’s output, which has the alpha channel we want to use, to the MatteControl node’s garbage matte input, which is a shortcut we can use to make a mask, matte, or alpha punch out a region of transparency in an image. Keep in mind that it’s easy to accidentally connect to the wrong input. Since inputs rearrange themselves depending on what’s connected and where the node is positioned, and frankly the colors can be hard to keep track of when you’re first learning, it’s key to make sure that you always check the tooltips associated with the input you’re dragging a connection over to make sure that you’re really connecting to the correct one. If you don’t, the effect won’t work, and if your effect isn’t working, the first thing you should always check is whether or not you’ve connected the proper inputs. One alternate method of connecting nodes together is to hold the Option key down while dragging a connection from one node’s output and dropping it onto the body of another node. This opens a pop-up menu from which you can choose the specific input you want to connect to, by name. Note that the menu only appears after you’ve dropped the connection on the node and released your pointing device’s button. Before/After Option-dragging a node connection to drop onto another node exposes a node input menu Once the Text1 node is properly connected to the MatteControl node’s Garbage Matte input, you should see a text-shaped area of transparency in the graphic if you load the MatteControl node into the Viewer. Chapter – 13 Learning to Work in the Fusion Page 2-114 (Top) Connecting the Text node to the Matte Control node’s garbage matte input, (Bottom) the resulting hole punched in the image Customizing Matte Control Nodes With this accomplished, we need to use the Inspector to change some parameters to get the result we want. In the Inspector controls for the Matte Control node, click the disclosure control for the Garbage Matte controls to expose their parameters. Because we actually have a Garbage matte connected, a variety of controls are available to modify how the garbage matte input is applied to the image. Click Invert to create the effect we really want, which is text that is textured with the chalkboard image. The alpha from the Text node punching a hole in another graphic However, the new chalkboard layer is far bigger than the HD-sized elements we’ve been working with, so the alpha from the Text1 node is too small. This is easily fixed by setting the Fit pop-up menu to “Width,” which automatically resizes the garbage matte to be as big as possible from edge to edge within the image. Chapter – 13 Learning to Work in the Fusion Page 2-115 The Garbage Matte settings of the MatteControl node Now, if we load the Merge1 node into the Viewer, we can see that the text effect is doing everything we want, but now the chalkboard text is too big. The final composite Using Transform Controls in the Merge Node Fortunately, there’s an easy fix that doesn’t even require us to add another node. Selecting the Merge1 node, we can see a set of transform parameters in the Inspector that specifically affect the foreground input’s image. This makes it quick and easy to adjust a foreground image to match the background. The final composite NOTE: When connecting two images of different sizes to a Merge node, the resolution of the background image defines the output resolution of that node. Keep that in mind when you run into resolution issues. Chapter – 13 Learning to Work in the Fusion Page 2-116 Dragging the Size slider to the left shrinks the text to create the effect we really want, and at this point, we’ve got the composite we need. The final composite Match Moving Text With Motion Tracking This next example introduces motion tracking, and how you can create a very simple match‑moving effect using the Tracker node, which is the Swiss army knife of trackers in the Fusion page. Adding a Layer We Want to Match Move In this example, we have a Text1 node that’s creating a “Switzerland” title, that’s composited over a drone shot flying over and around a mountain bridge. With the Text1 node selected, the on-screen controls that let you position the text it’s generating are visible in the Viewer, and the text is positioned where we’d like it to start. Note that, with the Text node selected, even the part of the text that’s off-screen can still be seen as an outline showing us where it is. Some text superimposed against a background, ready to track Chapter – 13 Learning to Work in the Fusion Page 2-117 Our goal for this composition is to motion track the background image so that the text moves along with the scene as the camera flies along. Setting Up to Track To set up for doing the motion track, we’ll begin by creating a disconnected Tracker node, using another method than those seen previously. Right-click anywhere in the background of the Node Editor (preferably where you want the new node to appear), and choose Add Tool > Tracking > Tracker from the contextual menu to create a new Tracker1 node underneath the MediaIn node. Creating a new node using the Node Editor contextual menu Next, we’ll drag a connection from the MediaIn1 node to the Tracker1 node to automatically connect the source clip to the Tracker1 background input. This branches the output from the MediaIn1 node to the Tracker node, so that the Tracker1 node processes the image separately from the rest of the node tree. This is not required, but it’s a nice organizational way to see that the Tracker node is doing an analysis that must be referred to in another way than a “physical” connection. Branching a Tracker node to use to analyze an image Chapter – 13 Learning to Work in the Fusion Page 2-118 A Simple Tracking Workflow The Tracker node is the simplest tracking operation the Fusion page has, and while there are several ways of using it. An extremely common workflow is to use the tracker node controls to analyze the motion of a subject in the frame with motion you want to follow, and then use the resulting motion path data by “connecting” it to the Center parameter of another node that’s capable of transforming the image you want to match move. Positioning the Tracker On-Screen Control When the Tracker node is selected, a single green box appears in the Viewer, which is the default on-screen control for the first default tracker that node contains (seen in the Tracker List of the Inspector controls). Keep in mind that you only see on-screen controls for nodes that are selected, so if you don’t see the on-screen tracker controls, you know you need to select the tracker you want to work with. Loading the tracker you want to work on into the Viewer is also the safest way to make sure you’re positioning the controls correctly relative to the actual image that you’re tracking. If you position your pointer over this box, the entire on-screen control for that tracker appears, and if you click the on-screen control to select that tracker, it turns red. As with so many other tracker interfaces you’ve likely used, this consists of two boxes with various handles for moving and resizing them: The inner box is the “pattern box,” which identifies the “pattern” in the image you’re tracking that you want to follow the motion of. The pattern box has a tiny handle at its upper-left-hand corner that you use to drag the box to overlap whatever you want to track. You can also resize this box by dragging any corner, or you can squish or stretch the box by dragging any edge, to make the box better fit the size of the pattern you’re trying to track. The center position of the tracker is indicated via x and y coordinates. The outer box is the “search box,” which identifies how much of the image the Tracker needs to analyze to follow the motion of the pattern. If you have a slow moving image, then the default search box size is probably fine. However, if you have a fast moving image, you may need to resize the search box (using the same kind of corner and side handles) to search a larger area, at the expense of a longer analysis. The name of that tracker is shown at the bottom right of the search box. The on-screen controls of a selected tracker seen in isolation It’s worth saying a second time, the handle for moving a tracker’s on-screen control is a tiny dot at the upper-left-hand corner of the inner pattern box. You must click on this dot to drag the tracker around. The handle for dragging the tracker boxes to move them around Chapter – 13 Learning to Work in the Fusion Page 2-119 In this example, we’ll drag the on-screen control so the pattern box overlaps a section of the bridge right over the leftmost support. As we drag the on-screen control, we see a zoomed-in representation of the part of the image we’re dragging over, to help us position the tracker with greater precision. For this example, the default sizes of the pattern and search box are fine as is. The zoomed-in preview that helps you position the pattern box as you drag it Using the Tracker’s Inspector Controls to Perform the Analysis At this point, let’s look at the Tracker node’s controls in the Inspector. There are a lot of controls, but for this simple example we only care about the main Trackers panel, with the tracking analysis buttons at the top, the tracking options below those, and the Tracker List underneath those. The Tracker list also has buttons for adding and deleting trackers, you have the option of adding multiple trackers that can be analyzed all at once for different workflows, but we don’t need that for now. Tracker Inspector controls, with the tracking analysis buttons at top, the tracker options in the middle, and the Tracker List below Chapter – 13 Learning to Work in the Fusion Page 2-120 Additional controls over each tracker and the image channels being analyzed appear at the bottom, along with offset controls for each tracker, but we don’t need those now (at least, not yet). Again, this track is so simple that we don’t need to change the default behaviors that much, but because the drone is flying in a circular pattern, the shape of the pattern area is changing as the clip plays. Fortunately, we can choose Every Frame from the Adaptive Mode pop-up, to instruct the tracker to update the pattern being matched at every frame of the analysis, to account for this. Changing the Adaptive Mode of the Tracker node to Every Frame to account for the camera’s shift of perspective Now, all we need to do is to use the tracker analysis buttons at top to begin the analysis. These buttons work like Transport controls, letting you start and stop analysis as necessary to deal with problem tracks in various ways. Keep in mind that the first and last buttons, Track From Last Frame and Track From First Frame, always begin a track at the last or first frame of the composition, regardless of the playhead’s current position, so make sure you’ve placed your tracker on-screen controls appropriately at the last or first frame. The analysis buttons, left to right, Track from last frame, track backward, stop tracking, track forward, track from first frame For now, clicking the Track From Beginning button will analyze the entire range of this clip, from the first frame to the last. A dialog lets you know when the analysis is completed, and clicking the OK button dismisses it so you can see the nice clean motion path that results. The analyzed motion path resulting from tracking a section of the bridge as the camera flies past Viewing Motion Track Data in the Spline Editor This is not a necessary part of the tracking workflow, but if you have an otherwise nice track with a few bumps in it, you can view the motion tracking data in the Spline Editor by viewing that tracker’s Displacement parameter curve. This curve is editable, so you can massage your tracking data in a variety of ways, if necessary. Chapter – 13 Learning to Work in the Fusion Page 2-121 Viewing motion tracking analysis data in the Spline Editor Connecting Motion Track Data to Match Move Now that we’ve got a successful analysis, it’s time to use it to create the Match Move effect. To make this process easier, we’ll double-click the tracker’s name in the Tracker list of the Inspector, and enter a new name that’s easier to keep track of (heh). Adding your own names make that tracker easier to find in subsequent contextual menus, and lets you keep track of which trackers are following which subjects as you work on increasingly complex compositions. Renaming a tracker to make it easier to find Now it’s time to connect the track we’ve just made to the text in order to start it in motion. Loading the Merge1 node into the Viewer to see the text in context with the overall composite we’re creating, we’ll select the Text1 node to open its parameters in the Inspector, and click the Layout panel icon (second button from the left) to expose the Layout controls, which are the text-specific transform controls used to position the text object in the Frame. These are the controls that are manipulated when you use the Text node on-screen controls for repositioning or rotating text. The Layout controls for a Text node, in the Layout panel Chapter – 13 Learning to Work in the Fusion Page 2-122 The Center X and Y parameters, while individually adjustable, also function as a single target for purposes of connecting to tracking to quickly set up match moving animation. You set this up via the contextual menu that appears when you right-click any parameter in the Inspector, which contains a variety of commands for adding keyframing, modifiers, expressions, and other automated methods of animation including connecting to motion tracking. If we right-click anywhere on the line of controls for Center X and Y, we can choose Connect To > Tracker1 > Bridge Track: Offset position from the contextual menu, which connects this parameter to the tracking data we analyzed earlier. Connecting the Center X and Y parameter to the “Bridge Track: Offset position” motion path we analyzed Immediately, the text moves so that the center position coincides with the center of the tracked motion path at that frame. This lets us know the center of the text is being match moved to the motion track path. The text now aligns with the motion track coordinate Offsetting the Position of a Match Moved Image In fact, we want to offset the match-moved text, so it’s higher up in the frame. To do this, we select the Tracker1 node again and use the Y Offset 1 dial control to move the text up, since now any changes we make to the Bridge Track dataset now apply to the center of the text that’s connected to it. Using the X and Y Offset controls in the Tracker1 node to offset the text layer’s position from the tracked motion path Chapter – 13 Learning to Work in the Fusion Page 2-123 The offset we create is shown as a dotted red line that lets us see the actual offset being created by the X and Y Offset controls. In fact, this is why we connected to the “Bridge Track: Offset position” option earlier. The text offset from the tracked motion path, the offset can be seen as a dotted red line in the Viewer Now, if we play through this clip, we can see the text moving along with the bridge. Two frames of the text being match moved to follow the bridge in the shot Chapter – 13 Learning to Work in the Fusion Page 2-124 Using Paint and Planar Tracking In this next example, we’ll take a look at a paint example in which we eliminate some facial scars on an actor’s forehead in a commercial. This workflow combines the Paint node with the Planar Tracking node, illustrating a common way of using these two powerful tools. The actor has some scars on his forehead that the director would like painted out Using a Planar Tracker to Steady a Subject to Paint Because this is a clip in motion, we can’t just paint out the scars on the man’s forehead, we need to deal with the motion so that the paint work we do stays put on his face. In this case, a common workflow is to analyze the motion in the image and use it to apply a “steady” operation, pinning down the area we want to paint in place so we can paint on an unmoving surface. The best way to do this in the Fusion page is to use the Planar Tracker, so we’ll add the PlanarTracker node after the MediaIn1 node such that the image we want to track is connected to the background input of the PlanarTracker node. As always, it’s important to be careful about which input you connect the image to for the effect to work properly. Adding a PlanarTracker node to analyze and steady the part of the image we want to paint on With the PlanarTracker node selected, and either it or the MediaOut1 node loaded in the Viewer, a Viewer toolbar appears with a variety of tools for drawing shapes and manipulating tracking data. The Planar Tracker works by tracking “planar” (read: flat) surfaces that you define by drawing a shape over the feature you want to track. When you first create a PlanarTracker node, you’re immediately put into a mode for drawing a shape, so in this case we draw a simple polygon over the man’s forehead, since that’s the feature we want to steady in preparation for painting. Chapter – 13 Learning to Work in the Fusion Page 2-125 We draw a simple box by clicking once each on each corner of the man’s forehead to create control points, clicking the first one we created to close the shape. Drawing a shape over the man’s forehead to prepare for Planar Tracking Turning our attention to the inspector, we can see that the Planar Tracker node has tracking transport controls that are similar to those of the Tracker, but with one difference. There are two buttons, Set and Go, underneath the Operation Mode pop-up, which defaults to “Track” since that’s the first thing we need to do. The Set button lets you choose which frame to use as the “reference frame” for tracking, so you should click the Set button first before clicking the Track Forward button below. Setting a reference frame at the beginning of the range of frames we want to track TIP: The Set button lets you supervise a Planar Track in progress and stop it if you see it slipping, making adjustments as necessary before clicking Set at the new frame to set a new reference before continuing to track forward towards the end of the clip. Chapter – 13 Learning to Work in the Fusion Page 2-126 The Pattern controls let you set up how you want to handle the analysis. Of these controls, the Motion Type pop-up menu is perhaps the most important. In this particular case, Perspective tracking is exactly the analysis we want, but in other situations you may find you get better results with the “Translation,” “Translation/Rotation,” and “Translation/Rotation/Scale” options that are available. Once you initiate the track, a series of dots appear within the track region shape you created to indicate trackable pixels found, and a green progress bar at the bottom of the Timeline Ruler lets you see how much of the shot is remaining to track. Clicking the Track From First Frame button to set the Planar Track in progress, green dots on the image and a green progress bar lets you know the track is happening NOTE: If you click one of the Track buttons to begin tracking and nothing happens, or you track for a few frames and then stop, that’s your cue that there isn’t enough trackable detail within the shape you’ve drawn for the Planar Tracker to work, and your best bet is to choose a different location of the image to track. Once the track is complete, you can set the Operation Mode of the PlanarTracker node’s controls in the Inspector to Steady. Setting the PlanarTracker node to Steady You’ll immediately see the image be warped as much as is necessary to pin the tracked region in place for whatever operation you want to perform. If you scrub through the clip, you should see that the image dynamically cornerpin warps as much as is necessary to keep the forehead region within the shape you drew pinned in place. In this case, this sets up the man’s head as a canvas for paint. Chapter – 13 Learning to Work in the Fusion Page 2-127 Steadying the image results in warping as the forehead is pinned in place for painting At this point, you’re ready to paint out those scars. Painting Over Blemishes Adding a Paint node after the PlanarTracker node gets us ready to paint. Adding a Paint node after the PlanarTracker to paint onto the steady surface With the Paint node selected and the MediaOut1 node loaded in the Viewer, we can see the paint tools in the Viewer toolbar. The first thing we want to do is to click on the fourth tool from the left, the “Stroke” tool, which is the preset tool for drawing strokes that last for the duration of the clip. The default “Multi-Stroke” tool is intended for frame by frame work such as painting out falling raindrops, moving dust and dirt, or other things of limited duration. The Stroke tool is much more appropriate when you want to paint out features or paint in fixes to subjects within the frame that need to remain in place for the whole shot. Choosing the Stroke tool from the Paint node’s tools in the Viewer toolbar Next, we need to go to the Inspector controls for the Paint node and choose the Clone mode from the Apply Controls. We’re going to clone part of the man’s face over the scars to get rid of them, and choosing the Clone mode switches the controls of the Paint node to those used for cloning. Chapter – 13 Learning to Work in the Fusion Page 2-128 Choosing the Clone mode in the Inspector There are additional controls located in this palette, however, that you should be familiar with. Brush Controls (at the top) contain the Brush Shape, Size, and Softness controls, as well as settings for how to map these parameters for tablet users. Apply Controls (in the middle) let you choose a paint mode, which includes Color, Clone, Emboss, Erase, Merge, Smear, Stamp, and Wire Removal. In this example we’ll be using Clone. The mode you choose updates what controls are available below. Stroke Controls (at the bottom) are intended to let you adjust strokes after they’ve been painted, and include controls for animating them with “write-on” effects, transforming strokes with standard sizing parameters, and adjusting brush spacing. With the Stroke tool selected in the Viewer tool bar, and Clone mode selected in the Inspector controls, we’re ready to start painting. If we move the pointer over the Viewer, a circle shows us the paint tool, ready to go. To use the clone brush, first you want to hold the Option key down and click somewhere on the image you want to clone from. In this example, we’ll sample from just below the first scar we want to paint. After Option-clicking to sample part of the image, clicking to begin painting sets an offset between where we’re sampling from and where we’re painting to, and dragging to draw paints a clone stroke. (Left) Setting an offset to sample for cloning, (Right) Dragging to draw a clone stroke Chapter – 13 Learning to Work in the Fusion Page 2-129 If you don’t like the stroke you’ve created, you can undo with Command-Z and try again. We repeat the process with the other scar on the man’s forehead, possibly adding a few other small strokes to make sure there are no noticeable edges, and in a few seconds we’ve taken care of the issue. (Top) Original image, (Bottom) After painting out two scars on the man’s forehead with the Stroke tool set to Clone TIP: You can adjust the size of the brush right in the Viewer, if necessary, by holding the Command key down and dragging the pointer left and right. You’ll see the brush outline change size as you do this. Before moving on, we’ll open the Modifiers panel of the Inspector, where we can see that every single paint stroke we’ve made appears as an item on the Modifiers list. This gives us access to the strokes we’ve painted for further modification. We don’t need to do anything at the moment, but when the time comes that you want to start making changes to strokes you’ve made, this is where they appear. Each stroke made appears as an entry with controls in the Modifiers panel of the Inspector Chapter – 13 Learning to Work in the Fusion Page 2-130 Keep in mind that the last stoke on the Modifiers list isn’t really a stroke, it’s a placeholder for the next stroke you’re about to make, which might explain the numbering of the strokes if you’re new to Fusion. Inverting the Steady Effect to Put the Motion Back In At this point, scrubbing through the clip shows that the paint strokes we’ve made are indeed sticking to the man’s forehead as we need them to do. Now we just have to invert the transform the Planar Tracker applied to put the clip back to the way it was, only with the painted fix attached in the process. This ends up being a two part process, but the first part is the simplest. Scrubbing through the steadied clip shows the paint fix is “sticking” to the man’s forehead Selecting and copying the PlanarTracker node coming before the Paint node, we select the Paint node and paste a copy of it after. This copy has all the analysis and tracking data of the original PlanarTracker node. Pasting a second copy of the PlanarTracker node after the Paint node With the second PlanarTracker node selected, we go into the Inspector and turn on the Invert Steady Transform checkbox, which in theory inverts the steady warp transform to put the image back to the way it was. However, in practice, the more the image needs to be warped to steady it, the more likely that inverting the warp will introduce other problems. Chapter – 13 Learning to Work in the Fusion Page 2-131 Turning on Invert Steady Transform to try and put the image back to the way it was. While the initial result appears to have just another warp applied to it, this time in reverse, the truth is that the region of the image centered on the shape used to do the planar analysis, the forehead, has gone back to the way it was before being steadied. It’s just the edges of the frame that are distorted. Using the Viewer’s Split Wipe Control This is a good example of a situation that can be tested using the Split Wipe control in the Viewer title bar. Opening the Split Wipe pop-up menu in the Viewer Using the Split Wipe pop-up, switch to B View (the current image is A View), then drag the second PlanarTracker node into the Viewer to load it into the B buffer, then switch back to A View and drag the MediaIn1 node into the Viewer to load it into the A buffer. Turning on the Split Wipe button displays a split screen of the original image (A) against the transformed image (B). You can drag the handle of the green split control to adjust the split, and you can drag the line to change the angle of the split (holding Shift lets you snap the angle to 45° angles). Comparing the “Invert Steady” version of the image with the original image to see the forehead is the same in both frames Chapter – 13 Learning to Work in the Fusion Page 2-132 So, the forehead is fine, but the rest of the image is now warping in an unusable way because of the extremity of the warp needed to steady the region we wanted to paint. That’s fine, because there’s an easy fix that’s a necessary part of this technique. Fixing the Edges by Only Using the Fixed Part of the Frame At this point, we’re ready for the second part of this fix, which is to mask and composite just the fixed forehead against the original clip. Isolating the Painted Forehead First, we need to mask out just the man’s painted forehead. We can do this by connecting a Polygon node to the garbage matte input of a MatteControl node, and then connecting the second PlanarTracker node’s output (with the fixed forehead) to the MatteControl node’s background input. This lets us draw a shape with the Polygon node and use it as a mask to crop out the man’s painted forehead. The placement of these two new nodes can be seen in the following screenshot. We can wire this up before drawing the shape, in fact it’s essential because otherwise you want to trace the image being fed to the MatteControl node using the Polygon node. Adding a Polygon node, a MatteControl node, and a Merge node to composite the painted forehead on the original clip TIP: When it comes to using Masks to create transparency, there are a variety of ways to do this, for example (a) attaching the image you want to mask to the background input of a Brightness/Contrast node with Alpha enabled to darken a hole in the alpha channel by lowering the Gain slider while the Polygon node is attached to the effect mask input, or (b) using ChannelBooleans to copy channel data to alpha from a Polygon node attached to the foreground input while the image you want to mask is attached to the background layer, however the MatteControl node is flexible enough and useful enough to merit learning about it now. Drawing a Polygon Mask Moving the playhead to the first frame of the clip, we’re ready to draw a mask to isolate the fixed forehead. Loading the MatteControl1 or MediaOut1 node into the Viewer, and selecting the Polygon1 node so that we see its tools in the Viewer toolbar sets us up for drawing a polygon. Drawing shapes using the Polygon node is similar to shape drawing in other spline-based environments, including the Color page: Clicking once draws a corner control point. Clicking and dragging creates a bezier curve. Click the first control point you created to close a shape. Chapter – 13 Learning to Work in the Fusion Page 2-133 We click and drag to create a shape that outlines the man’s forehead, and when we close the shape, we see exactly the opposite of what we want, a hole in the middle of the image. Drawing a shape to isolate the forehead gives an inverted result at first when using the Garbage Matte input of the MatteControl node to attach the Polygon to the MatteControl node Before fixing this, we drag the Soft Edge slider in the Inspector to the right to blur the edges just a bit. Inverting the Garbage Input Selecting the MatteControl1 node, we open the GarbageMatte controls, and click the Invert checkbox, which immediately gives us the result we want, of the forehead in isolation, ready for compositing. (Top) Inverting the Garbage Matte input, (Bottom) The resulting inverted mask inverting the forehead Compositing the Painted Forehead Against the Original Image Almost finished, we’ll add one more node, a Merge node, that we’ll use to actually layer the fixed forehead against the original image being output by the MediaIn node. Chapter – 13 Learning to Work in the Fusion Page 2-134 Creating a disconnected Merge node, we reconnect the MatteControl’s output to the green foreground input of the Merge node, and then pull out a second branch from the MediaIn1 node’s output to connect to the Merge node’s orange background input. This puts the cropped and fixed forehead on top of the original image. The painted forehead composited against the original image Match Moving the Mask to the Shot So now we’ve got the best of both worlds, a fixed forehead and the background of the shot looks good. However, if we select the Polygon node and then scrub forward in the clip, the fixed forehead mask drifts out of sync with the motion of the shot, so we have one last issue to deal with. Happily, match moving the mask to move with the shot is really simple. Because the Polygon isn’t animated to match the motion of the shot, it goes out of sync Selecting the first PlanarTracker node that comes right after the MediaIn node, and temporarily choosing Track from the Operation Mode pop-up menu, we can see there’s a Create Planar Transform button at the bottom of the listed controls. Clicking this button creates a new, disconnected node in the Node Editor that uses the planar track as a transform operation, for doing easy match moving. We click the Create Planar Transform button, and then set Operation Mode back to Steady. Creating a PlanarTransform node you can use to Match Move other images Chapter – 13 Learning to Work in the Fusion Page 2-135 We can insert this new node into the node tree to use it by holding the Shift key down and dragging it over the connection between the Polygon node and the MatteControl node, dropping it when the connection highlights. (Left) Inserting a PlanarTransform node by holding the Shift key down while dropping over a connection (Right) With the new PlanarTransform node inserted, the Polygon is automatically transformed to match the motion of the forehead that was tracked by the original PlanarTracker node, and it animates to follow along with the movement of the shot. At this point, we’re finished! The final painted image, along with the final node tree NOTE: While on-screen controls are only visible when you select the node they belong to, on-screen controls only appear transformed properly when you load a node downstream of the operations that would transform the image. Chapter – 13 Learning to Work in the Fusion Page 2-136 Building a Simple Green Screen Composite In this next example, we’ll take a look at how you can pre-organize the media you want to use in a composition in the Edit page, before creating a Fusion clip to bring all of it into the Fusion page in an organized way. Then, we’ll do a simple composite using a greenscreen key and two other layers to create a news story. Organizing Clips in the Edit Page to Create a Fusion Clip For this example, we’ll take a look at how you can organize multiple clips in the Edit page to use in the Fusion page by creating a “Fusion clip,” which is effectively a special-purpose compound clip used specifically by the Fusion page. The next effect we need to create involves a greenscreen clip, a background graphic, and a foreground graphic. This is the kind of situation where superimposing all three layers on the timeline to set up their order and timing can be the fastest way to set up the foundation of our composition. With these clips edited together, we select all of them, right-click the selection, and choose New Fusion Clip from the contextual menu. This embeds them all within a single clip, which is easy to manage in the Edit page and keeps all the relevant media necessary for this composition within one handy object. (Top) A stack of clips to use in a composite, (Bottom) Turning that stack into a Fusion clip in the Edit page Chapter – 13 Learning to Work in the Fusion Page 2-137 When we open the Fusion page to begin work, however, Fusion clips expose their contents in the Node Editor as a pre-built cascade of MediaIn nodes automatically connected by Merge nodes (one Merge node for each pair of clips) that take care of combining each layer of video the way they were in the Edit page timeline. The initial node tree of the three clips we turned into a Fusion clip With this node tree already assembled, we can focus our time on adding the nodes we’ll need to each branch of this tree, rather than assembling everything from scratch. Pulling a Greenscreen Key Using the Delta Keyer First, we’ll pull the greenscreen key we’ll need to create transparency behind the newscaster. To prepare, we’ll pull the Merge nodes off to the right to make room for the additional nodes we’ll be adding after the MediaIn nodes as we work. Creating space after the MediaIn nodes, and selecting the second one in preparation for adding a node Selecting the MediaIn2 node and loading the Merge1 node into the Viewer lets us see the greenscreen clip, and makes it easy for us to add a DeltaKeyer node inline by pressing ShiftSpace to open the Select Tool dialog with which to search for and insert any node. Adding a DeltaKeyer node inline after the MediaIn2 node The DeltaKeyer node is a sophisticated keyer that is capable of impressive results combining different kinds of mattes and a clean-plate layer, but it can also be used very simply if the background that needs to be keyed is well lit. And once the DeltaKeyer creates a key, it embeds the resulting alpha channel in its output, so in this simple case, it’s the only node we need to add. It’s also worth noting that, although we’re using the DeltaKeyer to key a green screen, it’s not limited to only keying green or blue; the DeltaKeyer can impressive keys on any color in your image. Chapter – 13 Learning to Work in the Fusion Page 2-138 With the DeltaKeyer selected, we’ll use the Inspector controls to pull our key, using a shortcut to quickly sample the shade of green from the background of the image. The shortcut we’ll use is a bit unorthodox, but it gives us the ability to preview how different areas of the background will key as we look for the right place to sample. We hold the Option key down and click the eyedropper tool, and while continuing to hold the Option key down, we drag the pointer over the green of the background in the Viewer. Option-clicking-and-dragging the eyedropper to the Viewer to sample the Background Color As we drag in the Viewer, an analysis of the color picked up by the location of the eyedropper appears within a floating tooltip, giving us some guidance as to which color we’re really picking. Meanwhile, we get an immediate preview of the transparency we’ll get at that pixel, and since we’re viewing the Merge1 node, this reveals the image we’ve connected to the background. (Before) The original image, (After) Sampling the green screen using the eyedropper from the Inspector When we’re happy with the preview, releasing the pointer button samples the color, and the Inspector controls update to display the value we’ve chosen. Chapter – 13 Learning to Work in the Fusion Page 2-139 The DeltaKeyer Inspector updates with the sampled color Now that we’ve selected a background color to pull a key with, we can load the DeltaKeyer node into the Viewer itself, and click the Color button (or select the Viewer and press C) to switch the viewer between the RGB color channels of the image and the Alpha channel, to evaluate the quality of the key. Loading the DeltaKeyer into the Viewer and clicking the Color button to view the alpha channel being produced A close examination of the alpha channel reveals some fringing in the white foreground of the mask. Happily, the DeltaKeyer has integrated controls for doing post-processing of the key being pulled, found in the third of the seven panels of controls available in the DeltaKeyer. Clicking the Matte panel opens up a variety of controls for manipulating the matte, and since the fringing we don’t like is on the foreground (white) part of the key, we’ll be using the Clean Foreground slider to make the fix. Chapter – 13 Learning to Work in the Fusion Page 2-140 Adjusting the Clean Foreground slider in the Matte panel of the DeltaKeyer controls In this case, raising the Clean Foreground slider a bit eliminates the inner fringing we don’t want, without compromising the edges of the key. (Before) The original key, (After) The key after using the Clean Foreground slider With this accomplished, we’re good with the key, so we load the Merge1 node back into the Viewer, and press C to set the Color control of the Viewer back to RGB. We can see the graphic in the background, but right now it’s too small to cover the whole frame, so we need to make another adjustment. The final key is good, but now we need to work on the background Chapter – 13 Learning to Work in the Fusion Page 2-141 Using the Transform Node to Resize a Background Since the background isn’t covering up the whole frame, we need to transform it. It’s a highresolution image, so that’s not a problem, however it’s connected to the background input of the Merge1 node, and although Merge nodes have built-in transform controls, they only work on the foreground input (on the premise that the foreground will need to be fit to the background). This means that we need to add a Transform node to the MediaIn1 node to take care of this. Selecting the MediaIn1 node and clicking the Transform button in the Toolbar takes care of this, and we’re ready to work. Adding a Transform node to change the sizing of the MediaIn1 image connected to the background While there are slider controls in the Inspector for Center, Size, and Angle (among other parameters), there are on-screen controls that give more satisfyingly direct control. Zooming out of the Viewer a bit by holding the Command key and using the scroll control of your pointer, we drag the side border of the graphic to proportionally enlarge the blue background until it fills the screen (there’s still a black border at the top and bottom of the clip, but that’s burned into the news clip we have). Enlarging the background to fill the frame using the Viewer’s on-screen controls At this point, we decide to make room for the graphic we know we’ll be putting into the frame at left, so we take advantage of the built-in transform controls in the Merge1 node that affect the foreground input. Selecting the Merge1 node, we drag the left arrow of the onscreen controls that appear to move the man to the right, and we take advantage of knowing the image of the man is high-resolution relative to our project resolution by dragging the side edge to proportionally enlarge the foreground image to crop out the black bars. Chapter – 13 Learning to Work in the Fusion Page 2-142 Using the Merge1 node’s on-screen transform controls to reposition and enlarge the image to prepare for adding another element NOTE: You may have noticed that there’s both Transform and Resize buttons in the Toolbar. It’s important to be aware that while the Transform node always refers to the original source resolution of the image for resolution-independent sizing in which multiple Transform nodes can scale the image down and up repeatedly with no loss of image resolution, the Resize node actually decreases image resolution when you shrink an image, or increases image resolution (with filtering) when enlarging. In most situations, you want to use the Transform node, unless you specifically want to alter and perhaps reduce image resolution to create a specific effect. Masking a Graphic Next, it’s time to work on the news graphic that will appear to the left of the man. If we load the Merge2 node, that combines the blue background and newscaster we just finished working on with the logo layer we brought into the Fusion page, we can see that the logo layer is actually a sheet of different logos that appear on top, so we need to cut one out using a mask and fit it into place. We need to mask out a single logo from this sheet to use in our composition Chapter – 13 Learning to Work in the Fusion Page 2-143 Selecting the MediaIn3 node that’s feeding the logo layer, we click the MatteControl button of the Toolbar to add a MatteControl node, and then we add a Rectangle mask, manually connecting the Rectangle mask’s output to the gray garbage mask input of the MatteControl node. Finally, we select the Rectangle node, and click it’s Invert checkbox to invert the Rectangle Mask’s output, so it’s cropping the logo layer correctly. Masking the logo using a Rectangle mask connected to a MatteControl node Now, all we need to do is to use the on-screen controls of the Rectangle mask to crop the logo we want to use, dragging the position of the mask using the center handle, and resizing it by dragging the top/bottom and left/right handles of the outer border. As an extra bonus, we can take care of the fact that the logo has rounded borders by using the Corner Radius slider in the Inspector controls for the Rectangle matte to add the same kind of rounding. Moving and resizing the mask to fit our logo, and rounding the edges using the Corner Radius Inspector control Chapter – 13 Learning to Work in the Fusion Page 2-144 Now that we’ve masked the logo, we’ll crop the unused parts of this image so that the logo we’re using is centered on the frame, which will make subsequent transform operations much easier. Selecting the MatteControl1 node, we add the Crop node from the Tools > Transform category of the Effects Library, and load the new node into the Viewer. Adding a Crop node after masking the image to center the cropped logo on the frame With the Crop node selected, we can click the crop tool in the Viewer toolbar Selecting the crop tool in the Viewer toolbar This lets us crop the image by dragging a bounding box around it. (Left) Dragging a bounding box using the Crop tool, (Right) The cropped logo now centered on the frame Chapter – 13 Learning to Work in the Fusion Page 2-145 NOTE: The Cropping node discards resolution, just like the Resize node does, so use it with care. At this point, we’re all set to move the logo into place, so we select the Merge2 node and load it into the Viewer, and once again avail ourselves of the built-in transform controls for foreground inputs, using the on-screen controls to put the logo where we want it and make it a suitable size. Placing the logo using the foreground input transform controls of the Merge2 node Animating an Image Using Keyframes We’re almost done with this grand tour of Fusion page functionality, but we’ve got one last task to accomplish. Now that we’ve positioned the logo appropriately, we need to animate it coming into frame to open the segment. To do this, we’ll use the keyframe controls in the inspector to begin keyframing, then we’ll use the controls in the Viewer to create a motion path, and finally we’ll use the Spline editor to refine the result. Animating a Parameter in the Inspector Before beginning to keyframe, it’s always good to think your way through what you want to do before starting anything, just to make sure you’re taking the right approach. In this case, we just want to slide the logo down from the top of the screen to where we’ve positioned it, so it’s probably best to start adding keyframes at the end point of the animation we want to create by moving the playhead in the Time Ruler 24 frames forward from the beginning of the composition. Selecting the Merge2 node, in which we used transform controls to position the logo, we click the small diamond control to the right of the Center parameter to create a keyframe for that parameter, in the process setting that parameter up so that every alteration we make on a different frame adds a keyframe. Adding a keyframe to begin animating a parameter Chapter – 13 Learning to Work in the Fusion Page 2-146 Next, we move the playhead back to the beginning of the composition, then zoom out of the Viewer so there’s more room around the frame before dragging the center handle of the logo up until we’ve dragged it off-screen. In the process, a second keyframe appears next to the Center parameter in the Inspector to show there’s a keyframe at that frame, and a motion path appears in the Viewer showing you the route the now animated logo will take. Moving an object in the Viewer to create animation via a motion path At this point, if we play though the animation, it’s functional, but not exciting. The motion is linear so it comes into the frame and stops with a nearly audible “thunk.” Happily, we can fix this using the Spline Editor. Using the Spline Editor Clicking the Spline button in the UI Toolbar opens the Spline Editor at the right of the Node Editor. The Spline Editor is a keyframe graph where you edit and finesse the curves created by animated parameters. By default, each animated parameter from every node in the current composition appears in the parameter list to the left of the curve graph. Turning on the Displacement checkbox shows our animated curve in the graph so we can work on it. The Displacement curve from the animated Center parameter of the Merge2 node in the Spline Editor Chapter – 13 Learning to Work in the Fusion Page 2-147 Drag a bounding box over the second of the two control points that are shown in the graph, so it’s highlighted. Selecting a control point to modify With that control point selected, click the Smooth button in the toolbar at the bottom of the Spline Editor to turn that keyframe into a bezier curve (this also works for multiple selected keyframes). This has the effect of easing the motion to a stop at that second keyframe. Clicking the Smooth button to turn the selected control point in the graph into a Bezier curve Playing through the animation, the logo does ease to a stop, but it’s subtle. We up the ante by dragging the bezier handle of the final keyframe to the left, making the curve steeper and resulting in the logo coasting to a stop more gradually. Editing the spline to create a steeper curve, making the logo coast more gradually to a stop Chapter – 13 Learning to Work in the Fusion Page 2-148 Congratulations At this point, we’re finished with our tour. As many things as we’ve covered, this is still only scratching the surface of what the Fusion page is capable of. However, this introduction should have given you a solid look at how to work in the Fusion page so that you can explore farther on your own. Have fun! Chapter – 13 Learning to Work in the Fusion Page 2-149 Chapter 14 Creating Fusion Templates The integration of Fusion into DaVinci Resolve 15 has resulted in one other exciting feature, and that’s the ability to use Fusion Titles in the Edit page. Fusion Titles are essentially generators that were created in the Fusion page, and which can be edited into the timeline of the Edit page as clips with custom controls. However, the really exciting this about this is that you can create your own Fusion Title templates, from nearly any Fusion page composition you can build using Fusion-generated objects such as Text+ layers, Fusion generators, and even 3D geometry and 3D text with texture and lighting effects. This brief chapter shows you how it’s done. Chapter – 14 Creating Fusion Templates 2-150 Contents Build a Template in the Fusion Page 2-152 Create a Macro 2-152 Restart Resolve and Use Your New Template 2-155 Chapter – 14 Creating Fusion Templates 2-151 Build a Template in the Fusion Page The first part of creating a Fusion template is to create a Fusion composition, consisting of Fusion-generated objects assembled to create nearly any kind of title or generator you can imagine. If you’re really ambitious, it can include animation. In this example, 3D titles and 2D titles have been combined into a show open. Building a composition to turn into a template Create a Macro Macros are basically Fusion compositions that have been turned into self-contained tools. Ordinarily, these tools are used as building blocks inside of the Fusion page so that you can turn frequently-made compositing tricks that you use all the time into your own nodes. However, we can also use this macro functionality to build templates for the Edit page. Having built your composition, select every single node you want to include in that template except for the MediaOut1 node. Selecting the nodes you want to turn into a Template Chapter – 14 Creating Fusion Templates 2-152 Having made this selection, right-click one of the selected nodes and choose Macro > Create Macro from the contextual menu. Creating a macro from the selected nodes The Macro Editor window appears, filled to the brim with a hierarchical list of every parameter in the composition you’ve just selected. The Macro Editor populated with the parameters of all the nodes you selected This list may look intimidating, but closing the disclosure control of the top Text1 node shows us what’s really going on. Chapter – 14 Creating Fusion Templates 2-153 A simple list of all the nodes we’ve selected Closing the top node’s parameters reveals a simple list of all the nodes we’ve selected. The Macro Editor is designed to let you choose which parameters you want to expose as custom editable controls for that macro. Whichever controls you choose will appear in the Inspector whenever you select that macro, or the node or clip that macro will become. So all we have to do now is to turn on the checkboxes of all the parameters we’d like to be able to customize. For this example, we’ll check the Text3D node’s Styled Text checkbox, the Cloth node’s Diffuse Color, Green, and Blue checkboxes, and the SpotLight node’s Z Rotation checkbox, so that only the middle word of the template is editable, but we can also change its color and tilt its lighting (making a “swing-on” effect possible). Turning on the checkboxes of parameters we’d like to edit when using this as a template Once we’ve turned on all the parameters we’d like to use in the eventual Template, we click the Close button, and a Save Macro As dialog appears. If we’re using macOS, we navigate to the / Library/Application Support/Blackmagic Design/DaVinci Resolve/Fusion/Templates/Edit/Titles directory, enter a name in the field below, and click Save. Chapter – 14 Creating Fusion Templates 2-154 Choosing where to save and name the Macro Restart Resolve and Use Your New Template After you’ve saved your macro, you’ll need to quit and re-open DaVinci Resolve. When you open the Effects Library of the Edit page, you should see your new template inside of the Titles category, ready to go in the Fusion Titles list. Custom titles appear in the Fusion Titles section of the Effects Library Chapter – 14 Creating Fusion Templates 2-155 Editing this template into the timeline and opening the Inspector, we can see the parameters we enabled for editing, and we can use these to customize the template for our own purposes. Customizing the template we made And that’s it! Chapter – 14 Creating Fusion Templates 2-156
Source Exif Data:
File Type : PDF File Type Extension : pdf MIME Type : application/pdf PDF Version : 1.6 Linearized : No Encryption : Standard V4.4 (128-bit) User Access : Print, Extract, Print high-res Create Date : 2018:04:19 10:27:43+10:00 Creator : Adobe InDesign CC 13.1 (Macintosh) Modify Date : 2018:04:19 10:35:31+10:00 Has XFA : No Language : en-US XMP Toolkit : Adobe XMP Core 5.6-c015 84.159810, 2016/09/10-02:41:30 Metadata Date : 2018:04:19 10:35:31+10:00 Creator Tool : Adobe InDesign CC 13.1 (Macintosh) Instance ID : uuid:7547b5f3-698c-0846-a6cf-9d0791ef727b Original Document ID : xmp.did:52a82e4f-fa56-48a1-aeac-e283b85f81d1 Document ID : xmp.id:c4dcb425-605c-4be3-90ef-022ddbb16a2d Rendition Class : proof:pdf Derived From Instance ID : xmp.iid:6dff42ee-fa1d-4246-8cf9-5d2da6a93a13 Derived From Document ID : xmp.did:01dfbdc6-0bd8-492c-a3a1-68b70fae458d Derived From Original Document ID: xmp.did:52a82e4f-fa56-48a1-aeac-e283b85f81d1 Derived From Rendition Class : default History Action : converted History Parameters : from application/x-indesign to application/pdf History Software Agent : Adobe InDesign CC 13.1 (Macintosh) History Changed : / History When : 2018:04:19 10:27:43+10:00 Format : application/pdf Producer : Adobe PDF Library 15.0 Trapped : False Page Layout : SinglePage Page Count : 282EXIF Metadata provided by EXIF.tools