Lytro poised to forever change filmmaking: debuts Cinema prototype and short film at NAB
|Lytro debuted its Cinema prototype to an eager crowd at NAB 2016 in Las Vegas, NV. It sports the highest resolution video sensor ever made.|
Lytro greeted a packed showroom at NAB 2016 in Las Vegas, Nevada to demo its prototype Lytro Cinema camera and platform, as well as debut footage shot on the system. To say we're impressed from what we saw would be an understatement: Lytro may be poised to change the face of cinema forever.
The short film 'Life', containing footage shot both on Lytro Cinema as well as an Arri Alexa, demonstrated some of the exciting applications of light field in video. Directed by Academy Award winner Robert Stromberg and shot by VRC Chief Imaging Scientist David Stump, 'Life' showcased the ability of light field to obviate green screens, allowing for extraction of backgrounds or other scene elements based off of depth information, and seamless integration of CGI elements into scenes. Lytro calls it 'depth screening', and the effect looked realistic to us.
Just as exciting was the demonstration of a movable virtual camera in post: since the light field contains multiple perspectives, a movie-maker can add in camera movement at the editing stage, despite using a static camera to shoot. And we're not talking about a simple pan left/right, up/down, or a simple Ken Burns effect... we're talking about actual perspective shifts. Up, down, left, right, back and forth, even short dolly movements - all simulated by moving a virtual camera in post, not by actually having to move the camera on set. To see the effect, have a look at our interview with Ariel Braunstein of Lytro, where he presents a camera fly-through from a single Lytro Illum shot (3:39 - 4:05):
The Lytro Cinema is capable of capturing these multiple perspectives because of 'sub-aperture imaging'. Head of Light Field Video Jon Karafin explains that in front of the sensor sits a microlens array consisting of millions of small lenses similar to what traditional cameras have. The difference, though, is that there is a 6x6 pixel array underneath each microlens, meaning that the image made up of only pixels on the sensor at any position (X,Y) underneath a microlens represents the scene as seen through one portion, or 'sub-aperture' of the lens. There will be 36 of these 'sub-aperture' images though, each providing one of 36 different perspectives, which then allows for computational reconstruction of the image with all the benefits of light field.
The 36 different perspectives affords you some freedom of movement in moving a virtual camera in post, but it is of course limited, affected by considerations like lens, focal length, and subject distance. It's not clear yet what that range of freedom is with the Cinema, but what we saw in the short film was impressive, something cinematographers will undoubtedly welcome in place of setting up motion rigs for small camera movements. Even from a consumer perspective, consider what auto-curation of user-generated content could do with tools like these. Think Animoto on steroids.
We've focused on depth screening and perspective shift, but let's not forget all the other benefits light field brings. The multiple perspectives captured mean you can generate 3D images or video from every shot at any desired parallax disparity (3D filmmakers often have to choose their disparity on-set, only able to optimize for one set of viewing conditions). You can focus your image after the fact, which saves critical focus and focus approach (its cadence) for post.* Selective depth-of-field is also available in post: you can choose whether you want shallow, or extended, depth-of-field, or even transition from selective to extensive depth-of-field in your timeline. You can even isolate shallow or extended depth-of-field to different objects in the scene using focus spread: say F5.6 for a face to get it all in focus, but F0.3 for the rest of the scene.
Speaking of F0.3 (yes, you read that right), light field allows you to simulate faster (and smaller) apertures previous thought impossible in post, which in turn places fewer demands on lens design. That's what allowed the Illum camera to house a 30-250mm equiv. F2.0 constant aperture lens in relatively small and lightweight body. You could open that aperture up to F1.0 in post, and at the demo of Cinema at NAB, Lytro impressed its audience with - we kid you not - F0.3 depth-of-field footage. A Lytro representative claimed even faster apertures can be simulated.
But all this doesn't come without a cost: the Lytro Cinema appears massive, and rightfully so. A 6x6 pixel array underneath each microlens means there are 36 pixels for every 1 pixel on a traditional camera; so to maintain spatial resolution, you need to grow your sensor, and your total number of pixels. Which is exactly what Lytro did - the sensor housing appeared to our eyes to be over a foot in width, sporting a whopping 755 million total pixels. That should mean that at worst, you'd get 755/36, or roughly 21MP final video output. Final output resolution was a concern with previous Lytro cameras: the Illum yielded roughly 5MP equivalent (sometimes worse) stills from a 40MP sensor. However, as we understand it, the theoretical lowest resolution of 21MP with the Cinema sensor means that output resolution shouldn't be a concern for 4K, or even higher-res, video.**
The optics appear as massive as the resolution, but that's partly because there are two optical paths: one for the 755MP light field capture, and the other to give the cinematographer a live preview for framing, focus, and exposure. The insane data rates for the light field capture, on the order of terabytes for every few seconds, means that Lytro Cinema comes with its own server on-set. The sensor is also actively cooled. The total unit lives on rails on wheels, so forget hand-held footage - for now. Bear in mind though, the original technicolor cinematic camera invented back in 1932 appeared similarly gargantuan, and Lytro specifically mentioned that different versions of Cinema are planned, some smaller in size.
Processing all that data isn't easy - in fact, no mortal laptop or desktop need apply. Lytro is partnering with Google to send footage to the cloud, where thousands of CPUs crunch the data and provide you real-time proxies for editing. Lytro stated the importance of integration with existing workflows, and to that end is building plug-ins to allow for light field video editing within existing editors - starting with Nuke. But Lytro is going a step further: they suggest the light field is the ultimate mastering format, and they're capable of converting all content - from footage to visual effects - into a 4D light field so you can, at any time, go back and re-render your film for any display device. This will be particularly important with the advent of holographic and other innovative light field displays.
The 4K footage from the Lytro Cinema that was mixed with Arri Alexa footage to create the short 'Life', viewed from our seating position, appeared comparable to what one might expect from professional cinema capture. CEO Jason Rosenthal commented that the short film was shot on both cameras to speak to how interchangeable footage can be with other cameras. Importantly, the footage appeared virtually noise free - which one might expect of such a large sensor area. Furthermore, Jon Karafin pointed out there are 'hundreds of input samples for every one output sample', which means a significant amount of noise averaging occurs, yielding a clean image, and a claimed 16 stops of dynamic range. In fact, in 'Life', noise had to be added back in to get the Lytro footage to match the Alexa.
That's incredibly impressive, given all the advantages light field brings. This may be the start of something incredibly transformative for the industry. After all, who wouldn't want the option for F0.3 depth-of-field with perfect focus in post, adjustable shutter angle and frame rate, compellingly real 3D imagery when paired with a light field display, and more? With increased capabilities for handling large data bandwidths, larger sensors, and more pixels, we think some form of light field will exist perhaps in most cameras of the future. Particularly when it comes to virtual reality capture, which Lytro also intends to disrupt with Immerge.
It's admirable just how far Lytro has come in such a short while, and we can't wait to see what's next. For more information, visit Lytro Cinema.
* If it's anything like the Illum, though, some level of focusing will still be required on set, as there are optimal planes of refocus-ability.
** We're not certain of the actual trade-off for the current Lytro Cinema. It's correlated to the number of pixels underneath each microlens, and effective resolution can vary at different focal planes, or change based on where focus was placed. This may be one reason for the overkill resolution - to ensure that at worst, capture is high resolution enough to meet high demands.
|Air to Air Refueling-9102 by vbuhay|
from Vehicle Refueling
|On the Catwalk by Lee8282|
|Yosemite Falls Midnight Reflection by Jonathan Shapiro|
from -Mirror in the Night Water- (Landscape in Full Colours Only)
Sony has updated its image sensor spec page and as expected, a few of the chips they make bear an uncanny resemblance to sensors found inside Fujifilm and Panasonic cameras.
This week Chris and Jordan are joined by renowned macro photographer Don Komarechka, who demonstrates a few simple techniques that can improve your macro photos in a big way.
The group that provides Canon users with programs to expand the feature set of their cameras has begun cracking the new EOS R mirrorless firmware.
The Pixel 3 represents another step forward in computational photography for Google's smartphone. We're just getting started with our testing – for now take a look at some sample images, including 'computational Raw' files available for download.
Lens Rentals Founder, Roger Cicala, has given the Canon EOS R one of his signature camera teardowns.
Nikon says firmware version 1.03 "Fixes an issue that in rare circumstances would delay the shutter release or the start of the autofocus operation."
The Kickstarter campaign for Yashica’s digiFilm Y35 camera has produced a wave of complaints about delays in shipping product as well as cameras that don’t work.
Pixelmator today released Pixelmator Pro 1.2 Quicksilver, a major update to its image editing app for Mac.
Although Raw performance of the EOS R is very similar to the 5D Mark IV, Canon's done some tweaking on the JPEGs - take a look at our studio scene to see for yourself.
If you've backed one of the company's crowdfunding projects, the reward will not arrive and you won't get your money back either as Meyer Optik Görlitz's parent company, Net SE, is completely dead.
The importance of APS-C, a future a7S model in development and why customers want two card slots – read our full interview with Sony's Kenji Tanaka.
Google's Super Res Zoom technology uses pixel-shifting methods to achieve zoom results comparable to some optical solutions. Google has published an in-depth explanation on its AI blog.
CyberLink has release the latest version of its photo editing and design program PhotoDirector.
Toy manufacturer Tomy has launched a no-battery-required smartphone printer that is remarkably like the one Holga has been promoting via a Kickstarter campaign but which is already available for $40/£39.
A handful of Sony users have noticed a particular model of SanDisk SD cards is showing errors when used with Sony a7 III camera.
The Fujifilm X-T3's 4K video more than lives up to its impressive specification, making it one of the most capable video cameras we've ever tested.
VSCO has made it easier to find the right presets for your photos with a few interface changes to its smartphone app.
TinyMOS is back with NANO1, an all-new astrophotography camera that's one-third the size of the TINY1 it announced three years ago.
Huawei's latest flagship device comes with the widest range of focal lengths of all current smartphones.
After shaking up the Lightroom ecosystem with Lightroom CC last year, Adobe has released version 2.0 of the cloud-centric photo organizer and editor. We look at new features like People View, how far Lightroom CC has come in its first year, and where Lightroom is headed.
Today, at Adobe MAX 2018, Adobe previewed Photoshop CC on iPad, a full-featured, desktop-class version of Photoshop for iOS.
The weather and has most definitely taken a turn toward fall here, and our shooting opportunities have followed suit. We brought the Canon RF 35mm F1.8 along to a harvest festival of sorts and a few of our usual haunts.
Pennsylvania Governor Tom Wolf has signed House Bill 1346 into effect, which imposes a fine upwards of $300 to drone operators who invade the privacy or harm the physical wellbeing of citizens.
Sigma is a company in flux, but CEO Kazuto Yamaki is undaunted by the upcoming prospect of developing lenses for eight lens mounts. The challenge will be keeping the company's identity along the way.
If you've been meaning to convert all of your old photos, video, and audio to digital formats, but simply lack the time or willpower to get through it all, a new service from Kodak will help you get the job done.
Almost all new cameras include impressive video features, but for the best results you'll often need an off-camera recorder. Chris and Jordan take a look at the brand new Ninja V from Atomos, and explain why it might just be one of the most useful tools you can add to your camera.
Collect allows you to transform 360-degree into a more easily digestible format by transforming it into directed traditional videos.
Sick of using your plain ol' keyboard to edit your photos in Lightroom and Photoshop? TourBox is hoping to expedite your post-production workflow using a clever combination of dials, buttons, and knobs.
Bag and accessory manufacturer Hex has launched two bags as part of its latest collection: the Clamshell Backpack and DSLR Sling.