Lytro poised to forever change filmmaking: debuts Cinema prototype and short film at NAB
|Lytro debuted its Cinema prototype to an eager crowd at NAB 2016 in Las Vegas, NV. It sports the highest resolution video sensor ever made.|
Lytro greeted a packed showroom at NAB 2016 in Las Vegas, Nevada to demo its prototype Lytro Cinema camera and platform, as well as debut footage shot on the system. To say we're impressed from what we saw would be an understatement: Lytro may be poised to change the face of cinema forever.
The short film 'Life', containing footage shot both on Lytro Cinema as well as an Arri Alexa, demonstrated some of the exciting applications of light field in video. Directed by Academy Award winner Robert Stromberg and shot by VRC Chief Imaging Scientist David Stump, 'Life' showcased the ability of light field to obviate green screens, allowing for extraction of backgrounds or other scene elements based off of depth information, and seamless integration of CGI elements into scenes. Lytro calls it 'depth screening', and the effect looked realistic to us.
Just as exciting was the demonstration of a movable virtual camera in post: since the light field contains multiple perspectives, a movie-maker can add in camera movement at the editing stage, despite using a static camera to shoot. And we're not talking about a simple pan left/right, up/down, or a simple Ken Burns effect... we're talking about actual perspective shifts. Up, down, left, right, back and forth, even short dolly movements - all simulated by moving a virtual camera in post, not by actually having to move the camera on set. To see the effect, have a look at our interview with Ariel Braunstein of Lytro, where he presents a camera fly-through from a single Lytro Illum shot (3:39 - 4:05):
The Lytro Cinema is capable of capturing these multiple perspectives because of 'sub-aperture imaging'. Head of Light Field Video Jon Karafin explains that in front of the sensor sits a microlens array consisting of millions of small lenses similar to what traditional cameras have. The difference, though, is that there is a 6x6 pixel array underneath each microlens, meaning that the image made up of only pixels on the sensor at any position (X,Y) underneath a microlens represents the scene as seen through one portion, or 'sub-aperture' of the lens. There will be 36 of these 'sub-aperture' images though, each providing one of 36 different perspectives, which then allows for computational reconstruction of the image with all the benefits of light field.
The 36 different perspectives affords you some freedom of movement in moving a virtual camera in post, but it is of course limited, affected by considerations like lens, focal length, and subject distance. It's not clear yet what that range of freedom is with the Cinema, but what we saw in the short film was impressive, something cinematographers will undoubtedly welcome in place of setting up motion rigs for small camera movements. Even from a consumer perspective, consider what auto-curation of user-generated content could do with tools like these. Think Animoto on steroids.
We've focused on depth screening and perspective shift, but let's not forget all the other benefits light field brings. The multiple perspectives captured mean you can generate 3D images or video from every shot at any desired parallax disparity (3D filmmakers often have to choose their disparity on-set, only able to optimize for one set of viewing conditions). You can focus your image after the fact, which saves critical focus and focus approach (its cadence) for post.* Selective depth-of-field is also available in post: you can choose whether you want shallow, or extended, depth-of-field, or even transition from selective to extensive depth-of-field in your timeline. You can even isolate shallow or extended depth-of-field to different objects in the scene using focus spread: say F5.6 for a face to get it all in focus, but F0.3 for the rest of the scene.
Speaking of F0.3 (yes, you read that right), light field allows you to simulate faster (and smaller) apertures previous thought impossible in post, which in turn places fewer demands on lens design. That's what allowed the Illum camera to house a 30-250mm equiv. F2.0 constant aperture lens in relatively small and lightweight body. You could open that aperture up to F1.0 in post, and at the demo of Cinema at NAB, Lytro impressed its audience with - we kid you not - F0.3 depth-of-field footage. A Lytro representative claimed even faster apertures can be simulated.
But all this doesn't come without a cost: the Lytro Cinema appears massive, and rightfully so. A 6x6 pixel array underneath each microlens means there are 36 pixels for every 1 pixel on a traditional camera; so to maintain spatial resolution, you need to grow your sensor, and your total number of pixels. Which is exactly what Lytro did - the sensor housing appeared to our eyes to be over a foot in width, sporting a whopping 755 million total pixels. That should mean that at worst, you'd get 755/36, or roughly 21MP final video output. Final output resolution was a concern with previous Lytro cameras: the Illum yielded roughly 5MP equivalent (sometimes worse) stills from a 40MP sensor. However, as we understand it, the theoretical lowest resolution of 21MP with the Cinema sensor means that output resolution shouldn't be a concern for 4K, or even higher-res, video.**
The optics appear as massive as the resolution, but that's partly because there are two optical paths: one for the 755MP light field capture, and the other to give the cinematographer a live preview for framing, focus, and exposure. The insane data rates for the light field capture, on the order of terabytes for every few seconds, means that Lytro Cinema comes with its own server on-set. The sensor is also actively cooled. The total unit lives on rails on wheels, so forget hand-held footage - for now. Bear in mind though, the original technicolor cinematic camera invented back in 1932 appeared similarly gargantuan, and Lytro specifically mentioned that different versions of Cinema are planned, some smaller in size.
Processing all that data isn't easy - in fact, no mortal laptop or desktop need apply. Lytro is partnering with Google to send footage to the cloud, where thousands of CPUs crunch the data and provide you real-time proxies for editing. Lytro stated the importance of integration with existing workflows, and to that end is building plug-ins to allow for light field video editing within existing editors - starting with Nuke. But Lytro is going a step further: they suggest the light field is the ultimate mastering format, and they're capable of converting all content - from footage to visual effects - into a 4D light field so you can, at any time, go back and re-render your film for any display device. This will be particularly important with the advent of holographic and other innovative light field displays.
The 4K footage from the Lytro Cinema that was mixed with Arri Alexa footage to create the short 'Life', viewed from our seating position, appeared comparable to what one might expect from professional cinema capture. CEO Jason Rosenthal commented that the short film was shot on both cameras to speak to how interchangeable footage can be with other cameras. Importantly, the footage appeared virtually noise free - which one might expect of such a large sensor area. Furthermore, Jon Karafin pointed out there are 'hundreds of input samples for every one output sample', which means a significant amount of noise averaging occurs, yielding a clean image, and a claimed 16 stops of dynamic range. In fact, in 'Life', noise had to be added back in to get the Lytro footage to match the Alexa.
That's incredibly impressive, given all the advantages light field brings. This may be the start of something incredibly transformative for the industry. After all, who wouldn't want the option for F0.3 depth-of-field with perfect focus in post, adjustable shutter angle and frame rate, compellingly real 3D imagery when paired with a light field display, and more? With increased capabilities for handling large data bandwidths, larger sensors, and more pixels, we think some form of light field will exist perhaps in most cameras of the future. Particularly when it comes to virtual reality capture, which Lytro also intends to disrupt with Immerge.
It's admirable just how far Lytro has come in such a short while, and we can't wait to see what's next. For more information, visit Lytro Cinema.
* If it's anything like the Illum, though, some level of focusing will still be required on set, as there are optimal planes of refocus-ability.
** We're not certain of the actual trade-off for the current Lytro Cinema. It's correlated to the number of pixels underneath each microlens, and effective resolution can vary at different focal planes, or change based on where focus was placed. This may be one reason for the overkill resolution - to ensure that at worst, capture is high resolution enough to meet high demands.
Sigma says its 70-200mm F2.8 DG OS HSM Sport lens is set to hit shelves by the end of December 2018 at a retail price of $1,499.
DxO PhotoLab 2.1 brings a collection of new features to MacOS and Windows users alike.
The new 'Elegant' lens series includes entirely manual F2.4 lenses in 24mm, 35mm, 50mm, 75mm and 90mm focal lengths.
A feature alerts pilots visually and/or verbally when their drone is approaching airspace that is unsafe or areas where drone flying is not permitted.
GoPro announced Monday morning that it plans to move production of United States-bound cameras out of China, citing tariffs concerns.
The Sigma 56mm F1.4 combines a sensible sub-$500 price tag and excellent performance, providing a portrait-friendly 85mm equiv. view on Sony's APS-C mirrorless cameras.
Azriel Knight of the YouTube channel This Old Camera explains the history of DX encoding.
The 250mm F4 is Fujifilm's longest lens for its medium-format system. It's equivalent to about 200mm on a GFX camera, and we put it to work on some portraits as well as some scenes around Seattle's waterfront – take a look.
Sony has removed the ability to download firmware version 2.0 for its a7 III and a7R III mirrorless cameras from its website.
Handing out awards for the best gear of the year is a big job, so we called in some reinforcements from Calgary to help us.
A new patent from Canon lays out the schematics for a speedbooster-style adapter for mounting Canon EF lenses onto EOS M cameras, but with a variable baffle to reduce the risk of flare.
The Jackson Hole Travel and Tourism Board has started a campaign asking visitors to stop geotagging their specific locations when visiting Wyoming's national parks.
Film simulation app Filmborn has been updated with new presets, features, and overall improved support on Apple's latest mobile operating system and devices.
The Colorado Tripod Company has introduced what it claims is the world’s first titanium tripod system, with a funding campaign on Kickstarter.
We've been shooting with the LX100 II both in and out of the studio, as part of our ongoing review. We're pretty impressed, so far, with the revised JPEG color and addition of a touchscreen both noticeable improvements.
An upcoming Xiaomi smartphone might use a 48MP sensor for pixel-binning, high-quality digital zooming and other algorithm-powered imaging features.
It's not cheap, but you may soon be able to get your hands on peel apart film once again thanks to ONE INSTANT.
Skylum's Luminar 3 arrives on December 18 with the long-awaited ability to manage your photo library. However, it won't be a full DAM (digital asset manager); the company plans to roll out features throughout 2019 and won't charge for updates from Luminar 2018 during that time.
Hasselblad has released an update to its Phocus post-production software that brings new and updated tools, as well as updated native lens support.
Nikon's IPTC Preset Manager, a tool for creating predetermined sets of metadata, has received an update. Version 1.1.0 no longer uses Microsoft Silverlight, sheds the network connection requirement, adds extended language support, updates support for Windows 8.1 and Windows 10, and ends support for Windows Vista and Windows XP.
Insta360 has launched a software update for its One X 360-degree camera and announced a camera bundle exclusively available on Apple.com.
Xiaomi has laid out the details for its new AI-powered image processing platform DeepExposure.
The Qualcomm Snapdragon 855 chipset is expected to power most 2019 high-end Android phones, including the Samsung Galaxy S10.
Camera app developer Hipstamatic says it has found a way to use the depth data generated by the iPhone X to improve the way its TinType app works out which areas of a picture to render out of focus.
If you're looking for a high-quality camera, you don't need to spend a ton of cash, nor do you need to buy the latest and greatest new product on the market. In our latest buying guide we've selected some cameras that while they're a bit older, still offer a lot of bang for the buck.
The Verizon-owned social network platform Tumblr has announced it will be removing all adult content - including photos - from its platform starting December 17th, 2018.
Guests who would rather spend time actually enjoying their Swiss vacation can now do so while still maintaining a presence on social networks, thanks to Ibis Switzerland Hotels' new social media sitting services.
Two challenges to Apple’s claim that its iPhone X can shoot studio quality portraits have been turned down by the UK’s Advertising Standards Authority (ASA).
We take a head-to-head look at the Apple iPhone Xs's bokeh effect versus a 58mm Nikkor lens on full-frame. The results? Well, we're pretty impressed.