Digital Media Net - Your Gateway To Digital media Creation. News and information on Digital Video, VR, Animation, Visual Effects, Mac Based media. Post Production, CAD, Sound and Music

Assimilate Announces Open Beta of New Live FX for Live Compositing of Virtual Productions

Live Compositing Enables Real-Time, On-Set Pre-Visualization of Final Content

October 19, 2021 – Assimilate – San Jose, CA – Assimilate® announced today an open beta for its new and innovative Live FX® software that enables real-time, live compositing for green-screen and LED wall-based virtual productions on set. Live FX is also ideal for creating quick comps when scouting locations. Live FX features the latest technology in keying, camera tracking, DMX light control, Notch Block integration, and a live link to Unreal Engine that simplifies on-set virtual production for not only previsualization, but even final pixel for in-camera VFX. By recording any incoming dynamic metadata along with the imagery, Live FX can automatically prep all content for VFX/post, saving time and money. Filmmakers and artists – from freelancers to big studios – can now work in a familiar film/video workflow rather than in a programming style environment.

Control any parameter in Live FX using Open Sound Control with an iPad

“The future of filmmaking is in virtual productions for the obvious reasons of time and cost savings in reduced logistics, travel, equipment, shipping, location set-ups, and more,” said Jeff Edson, CEO at Assimilate. “Assimilate is supporting this process by pioneering live compositing on set, which provides a significant leap in productivity and creativity for users. Live FX enables the most realistic pre-visualization up front in the workflow before the content moves down the pipeline and into post production. Additionally, we wanted to develop a tool that meets both needs of running on laptops for doing quick comps for location scouting or on-set quick checks, as well as on powerful systems to drive entire productions.”

Camera Tracking

Live FX powerful HSV-Keyer includes camera tracking to create final pixel in-camera

Live FX supports a broad range of tracking systems for a variety of on-set situations, such as indoor, outdoor, big stage or small set, tethered or untethered. To track camera movement, Live FX allows use of Intel RealSense tracking cameras, Mo-Sys StarTracker, HTC Vive trackers, and even gyro and ARKit technology inside smartphones mounted to the cinema camera. Even without a dedicated tracking device, Live FX is able to create camera tracks using its own live tracking algorithm. All tracking data can be refined and synced to the incoming live camera signal, so all data stays in sync throughout the entire pipeline. Through its built-in virtual camera calibration, Live FX is able to calculate lens distortion and create final composites live on-set.

Camera tracking in Live FX: Supports Mo-Sys StarTracker, Intel RealSense, HTC Vive, and even smartphones via gyro-sensor and ARKit implementations

Easy Green Screen Replacement

Live FX on-set: Preparing a live green-screen composite

The green screen can be easily replaced on set with a virtual background of choice – whether it’s a simple 2D texture, 360 equirectangular footage, a 3D Notch Block, or a live feed from another camera or Unreal Engine. This is made possible by a variety of keying algorithms within Live FX. Users can create and combine multiple keyers using the layer stack and add multiple garbage masks to their composite. RGB and Alpha channels can be output separately via SDI or NDI to be captured by other tools on-set, such as Unreal Engine.

LED Rear Projection Creates Photorealistic Environment

Live FX supports color-managed playback of any footage, at any resolution, and at any frame rate.
It allows projection of conventional 2D files, 360 equirectangular footage, Notch Blocks, or live feeds, such as 3D environments directly from Unreal Engine, onto LED walls. Once the content is loaded, it can be color graded, framed and tweaked further.

DMX Lighting Control

Live FX control of studio lighting via DMX, based on back-projected video content for a car-chase scene

For a realistic and natural look, accurate lighting is a key aspect of virtual productions. The DMX lighting control engine in Live FX allows users to mark a region of interest on the footage that is being projected onto the LED wall. Live FX analyzes and averages the color inside the marked region for every frame and sends the color down to the stage lighting, making it easy and straight forward to adjust the lighting of any scene.  Think big – not only bright blue sky, but even dynamic scenes like a house on fire with flickering ARRI Sky panels or a car-chase scene through tunnels are now easy to create and set up.

Live Links

Live FX features Live Links, which allows any piece of dynamic metadata entering the system to be used inside a live composition. The metadata comes in from a dedicated tracker system, via the SDI signal of the camera, or through the Open Sound Control (OSC) protocol from another system. Examples include using the dynamic lens metadata that comes in from the camera through SDI to directly animate the blur of a texture. Or, link the camera’s own gyro metadata to the virtual camera inside Live FX to link camera movement completely without additional tracking devices.

Timecode Sync Signal Routing

One of the most important aspects in live compositing is managing latency. Live FX allows for compositing multiple streams together: Live streams, dynamic renders like those from Unreal Engine, and any pre-recorded footage. If an input is timecode-managed, Live FX can auto sync multiple feeds by delaying the fastest feeds to ensure 100 percent frame sync in the composite. The camera tracking can be manually delayed to more precisely align the tracking data with the video feed. The additional latency from SDI output is kept to a minimum in Live FX and the output can be genlocked. Users can also exchange image data using NDI, or even direct GPU texture sharing with zero latency.

Recording, Metadata, and Post Prep

Anything that goes into Live FX can be recorded and instantly played back for review. The media can be recorded in H264, Avid DNx or even Apple ProRes, including alpha channel. In the case of multiple live feeds, users can decide to record the final baked result or the individual raw feeds separately. And not only the media is recorded; all metadata that is captured and used in the live composite – including camera tracking data – is saved to file as well. This all allows for a unique and highly efficient workflow: Start with the live composite and record the raw feeds and metadata separately. As soon as the recording stops, Live FX automatically creates an offline composite that is ready for instant playback and review. The offline composite is a duplicate of the live composite, using the just recorded files instead of the live signals, and includes all other elements and animation channels for further manipulation.

Once the high resolution media is offloaded from the camera, Live FX automatically assembles the online composite and replaces the on-set recorded clips with the high quality camera raw media.

These online composites can be loaded into Assimilate’s Scratch® to create metadata-rich dailies or VFX plates in the OpenEXR format, including all the recorded frame-based metadata, which saves time in post-production.

“Live FX is not an island. It’s part of Assimilate’s extensive Scratch ecosystem of advanced digital-cinema tools that are designed not only to simplify workflows and increase productivity for DITs, post-production artists, directors, and all filmmakers, but also to achieve the highest quality productions,” said Mazze Aderhold, chief strategic workflow engineer at Assimilate. “For Live FX, Assimilate is now expanding its connections to third-party software with an Unreal Engine plug-in for metadata communication, support for Open Sound Control (OSC) protocol, and direct GPU texture sharing to enable ultimate flexibility on-set. We’ve also added support for a wide range of different camera tracking devices and workflows that equally serve high-end productions and indie filmmakers with cost-effective solutions for virtual production – all running on off-the-shelf (OTF) hardware. This makes Live FX the easy go-to solution for all virtual productions.”

Participate in Open Beta

To participate in the open beta, simply visit the Assimilate webpage and click the “Try” button for Live FX. Fill in your email address and receive a return email with the download-link and license key.

For beta users to get started immediately, watch the Live FX short hands-on video tutorial series.

www.assimilateinc.com

About Assimilate

Assimilate develops and offers advanced, real-time tools for on-set and post-production workflows that deliver groundbreaking features, speed, performance, flexibility, and stability for industry professionals. The recently launched Live FX is a state-of-the-art live compositor, shipping with camera tracking, DMX control and the latest green-screen keying technology for virtual production. Live Assist is the multi-cam VTR tool for video assist, offering support for any number of cameras, any resolution, with easy green-screen comp and local-clip server functionality. Live Looks is the optimum tool for live grading and direct-look creation with an instant connection to Assimilate’s Scratchâ  for live streaming and keying capabilities. Used by DITs and post artists worldwide, Scratch is known for its advanced on-set and post-production tools that offer stability, speed and metadata support for any camera format, including 2D/3D/VR. Play Pro is the go-to solution for a professional player for VFX-reviews, ProRes RAW QC and genuine ProRes transcoding on Windows.

Staff

Recent Posts

Vision Products, LLC to Demonstrate Zero Distortion HMD at I/ITSEC 2024 in Booth 524

CAMPBELL, Calif.--(BUSINESS WIRE)--Vision Products, LLC, a leader in head/helmet mounted displays (HMDs) for military training…

3 hours ago

GRAMMY Museum® Expands GRAMMY Camp® to New York and Miami Beginning Summer 2025, Providing Greater Access to Aspiring Young Musicians and Creative Entrepreneurs Nationwide

LOS ANGELES--(BUSINESS WIRE)--The GRAMMY Museum® announced today the expansion of its acclaimed GRAMMY Camp® ,…

3 hours ago

Metinteractive Supports New Atrium Health Amphitheater in Macon, Georgia

Metinteractive, which provides strategic solutions for architecture, communication and technology, drew from its extensive experience…

4 hours ago

Mega Matrix Inc. Signs MOU with 9Yards Cinema Production to Establish A $100m Investment Fund for Global Short Drama and Pan-Entertainment Sector

SINGAPORE and ABU DHABI, UAE, Nov. 21, 2024 /PRNewswire/ -- Mega Matrix Inc. (NYSE American: MPU) today…

11 hours ago

Wondershare Filmora, Universal Music for Creators Team to Empower Video Editors with World-Class Recordings

VANCOUVER, BC, Nov. 21, 2024 /PRNewswire/ -- Wondershare Filmora, a leading video editing software, and…

11 hours ago

LetinAR Unveils First Mass-Produced AR Product Through Partnership with Japan’s Largest Telecom Group

SAN JOSE, Calif., Nov. 21, 2024 /PRNewswire/ -- LetinAR, a leader in optical technology for…

11 hours ago