zLense declares ‘world first’ in 3D camera technology27 October 2014
zLense has announced the launch of the world’s first depth-mapping camera solution that captures 3D data and scenery in realtime and adds a 3D layer to the footage.
The technology processes space information, making new and real three-dimensional compositing methods possible, enabling production teams to create 3D effects and utilise state-of-the-art CGI in live TV or pre-recorded transmissions, with no special studio set up.
Utilising the solution, directors can produce unique simulated and augmented reality worlds, generating and combining dynamic virtual reality and augmented effects in live studio or outside broadcast transmissions. The depth-sensing technology allows for 360° of camera movement, and directors can combine dolly, jib arm and handheld shots as presenters move within, interact with and control the virtual environment and, in the near future, do so using only motions.
“We’re poised to shake up the virtual studio world by putting affordable high-quality realtime CGI into the hands of broadcasters,” said Bruno Gyorgy, president of zLense. “This unique world-leading technology changes the face of TV broadcasting as we know it, giving producers and programme directors access to CGI tools and techniques that transform the audience viewing experience.”
The VFX realtime capabilities enabled by the zLense Virtual Production platform include volumetric effects; shadows and reflections to create convincing visual appearances; 3D distortions; creation of a virtual environment with interactive physical particle simulation; wide shot and in-depth compositions with full body figures; and realtime Z-map and 3D models of the picture.
The platform combines depth-sensing technology and image-processing in a standalone camera rig that captures the 3D scene and camera movement. The ‘matte box’ sensor unit removes the need for external tracking devices or markers, and can be mounted on most camera rigs. The zLense Virtual Production platform can be used alongside other, pre-exisiting, rendering engines, VR systems and tracking technologies.