Lytro poised to forever change filmmaking: debuts Cinema prototype and short film at NAB
|Lytro debuted its Cinema prototype to an eager crowd at NAB 2016 in Las Vegas, NV. It sports the highest resolution video sensor ever made.|
Lytro greeted a packed showroom at NAB 2016 in Las Vegas, Nevada to demo its prototype Lytro Cinema camera and platform, as well as debut footage shot on the system. To say we're impressed from what we saw would be an understatement: Lytro may be poised to change the face of cinema forever.
The short film 'Life', containing footage shot both on Lytro Cinema as well as an Arri Alexa, demonstrated some of the exciting applications of light field in video. Directed by Academy Award winner Robert Stromberg and shot by VRC Chief Imaging Scientist David Stump, 'Life' showcased the ability of light field to obviate green screens, allowing for extraction of backgrounds or other scene elements based off of depth information, and seamless integration of CGI elements into scenes. Lytro calls it 'depth screening', and the effect looked realistic to us.
Just as exciting was the demonstration of a movable virtual camera in post: since the light field contains multiple perspectives, a movie-maker can add in camera movement at the editing stage, despite using a static camera to shoot. And we're not talking about a simple pan left/right, up/down, or a simple Ken Burns effect... we're talking about actual perspective shifts. Up, down, left, right, back and forth, even short dolly movements - all simulated by moving a virtual camera in post, not by actually having to move the camera on set. To see the effect, have a look at our interview with Ariel Braunstein of Lytro, where he presents a camera fly-through from a single Lytro Illum shot (3:39 - 4:05):
The Lytro Cinema is capable of capturing these multiple perspectives because of 'sub-aperture imaging'. Head of Light Field Video Jon Karafin explains that in front of the sensor sits a microlens array consisting of millions of small lenses similar to what traditional cameras have. The difference, though, is that there is a 6x6 pixel array underneath each microlens, meaning that the image made up of only pixels on the sensor at any position (X,Y) underneath a microlens represents the scene as seen through one portion, or 'sub-aperture' of the lens. There will be 36 of these 'sub-aperture' images though, each providing one of 36 different perspectives, which then allows for computational reconstruction of the image with all the benefits of light field.
The 36 different perspectives affords you some freedom of movement in moving a virtual camera in post, but it is of course limited, affected by considerations like lens, focal length, and subject distance. It's not clear yet what that range of freedom is with the Cinema, but what we saw in the short film was impressive, something cinematographers will undoubtedly welcome in place of setting up motion rigs for small camera movements. Even from a consumer perspective, consider what auto-curation of user-generated content could do with tools like these. Think Animoto on steroids.
We've focused on depth screening and perspective shift, but let's not forget all the other benefits light field brings. The multiple perspectives captured mean you can generate 3D images or video from every shot at any desired parallax disparity (3D filmmakers often have to choose their disparity on-set, only able to optimize for one set of viewing conditions). You can focus your image after the fact, which saves critical focus and focus approach (its cadence) for post.* Selective depth-of-field is also available in post: you can choose whether you want shallow, or extended, depth-of-field, or even transition from selective to extensive depth-of-field in your timeline. You can even isolate shallow or extended depth-of-field to different objects in the scene using focus spread: say F5.6 for a face to get it all in focus, but F0.3 for the rest of the scene.
Speaking of F0.3 (yes, you read that right), light field allows you to simulate faster (and smaller) apertures previous thought impossible in post, which in turn places fewer demands on lens design. That's what allowed the Illum camera to house a 30-250mm equiv. F2.0 constant aperture lens in relatively small and lightweight body. You could open that aperture up to F1.0 in post, and at the demo of Cinema at NAB, Lytro impressed its audience with - we kid you not - F0.3 depth-of-field footage. A Lytro representative claimed even faster apertures can be simulated.
But all this doesn't come without a cost: the Lytro Cinema appears massive, and rightfully so. A 6x6 pixel array underneath each microlens means there are 36 pixels for every 1 pixel on a traditional camera; so to maintain spatial resolution, you need to grow your sensor, and your total number of pixels. Which is exactly what Lytro did - the sensor housing appeared to our eyes to be over a foot in width, sporting a whopping 755 million total pixels. That should mean that at worst, you'd get 755/36, or roughly 21MP final video output. Final output resolution was a concern with previous Lytro cameras: the Illum yielded roughly 5MP equivalent (sometimes worse) stills from a 40MP sensor. However, as we understand it, the theoretical lowest resolution of 21MP with the Cinema sensor means that output resolution shouldn't be a concern for 4K, or even higher-res, video.**
The optics appear as massive as the resolution, but that's partly because there are two optical paths: one for the 755MP light field capture, and the other to give the cinematographer a live preview for framing, focus, and exposure. The insane data rates for the light field capture, on the order of terabytes for every few seconds, means that Lytro Cinema comes with its own server on-set. The sensor is also actively cooled. The total unit lives on rails on wheels, so forget hand-held footage - for now. Bear in mind though, the original technicolor cinematic camera invented back in 1932 appeared similarly gargantuan, and Lytro specifically mentioned that different versions of Cinema are planned, some smaller in size.
Processing all that data isn't easy - in fact, no mortal laptop or desktop need apply. Lytro is partnering with Google to send footage to the cloud, where thousands of CPUs crunch the data and provide you real-time proxies for editing. Lytro stated the importance of integration with existing workflows, and to that end is building plug-ins to allow for light field video editing within existing editors - starting with Nuke. But Lytro is going a step further: they suggest the light field is the ultimate mastering format, and they're capable of converting all content - from footage to visual effects - into a 4D light field so you can, at any time, go back and re-render your film for any display device. This will be particularly important with the advent of holographic and other innovative light field displays.
The 4K footage from the Lytro Cinema that was mixed with Arri Alexa footage to create the short 'Life', viewed from our seating position, appeared comparable to what one might expect from professional cinema capture. CEO Jason Rosenthal commented that the short film was shot on both cameras to speak to how interchangeable footage can be with other cameras. Importantly, the footage appeared virtually noise free - which one might expect of such a large sensor area. Furthermore, Jon Karafin pointed out there are 'hundreds of input samples for every one output sample', which means a significant amount of noise averaging occurs, yielding a clean image, and a claimed 16 stops of dynamic range. In fact, in 'Life', noise had to be added back in to get the Lytro footage to match the Alexa.
That's incredibly impressive, given all the advantages light field brings. This may be the start of something incredibly transformative for the industry. After all, who wouldn't want the option for F0.3 depth-of-field with perfect focus in post, adjustable shutter angle and frame rate, compellingly real 3D imagery when paired with a light field display, and more? With increased capabilities for handling large data bandwidths, larger sensors, and more pixels, we think some form of light field will exist perhaps in most cameras of the future. Particularly when it comes to virtual reality capture, which Lytro also intends to disrupt with Immerge.
It's admirable just how far Lytro has come in such a short while, and we can't wait to see what's next. For more information, visit Lytro Cinema.
* If it's anything like the Illum, though, some level of focusing will still be required on set, as there are optimal planes of refocus-ability.
** We're not certain of the actual trade-off for the current Lytro Cinema. It's correlated to the number of pixels underneath each microlens, and effective resolution can vary at different focal planes, or change based on where focus was placed. This may be one reason for the overkill resolution - to ensure that at worst, capture is high resolution enough to meet high demands.
Jun 5, 2017
Jun 20, 2017
Jun 3, 2017
Jun 2, 2017
|Hot Air Balloons Over Bagan by User9320321874|
|Yellow Warbler by LeeS|
from A Big Year - birds
|Waiting for the Parade by tcoker1103|
from - La Vida Loca - (Black and White Street Photography+ A Border)
If you're thinking of using Canon's sports glass on the Sony a9, think again. The ultra-fast camera slows way down when you attach off-brand glass.
The Polish town of Katowice is not known as an area of beauty, but as all photographers know, that doesn't mean that beauty can't be found if you know where to look. Mariusz Pietranek used a drone to look down on the colorful sedimentation tanks at an ironworks.
New York Times video journalist Ben Solomon spent a harrowing three weeks accompanying Iraqi Major Sajjad al-Hour as he and his men fought to retake Mosul from I.S. forces.
The 3D VR camera launched through a crowdfunding campaign in 2015 goes on sale beginning June 26.
Noctilucent clouds, a crescent moon and Venus were visible in the pre-dawn sky over Budapest yesterday. Photographer György Soponyai captured NASA's astronomy picture of the day.
Squirming pets won't sit still for photos? A Kickstarter campaign is looking to help.
Find out how Chris Burkard shifted from editorial photography to his true passions: landscapes, conservation and, of course, surfing.
The updated EyeEm app scans your camera roll and picks images that are composed particularly well, have the best quality, or highest chance of selling on EyeEm Market.
It's three years old but still a solid option for a Micro Four Thirds shooter looking for a high-quality, fast, wide-angle prime. Take a look at how we got along with it.
Tamron has announced the longest all-in-one zoom lens currently available, the 18-400mm F3.5-6.3 Di II VC HLD. Designed for Canon and Nikon crop-sensor cameras, the lens will be available in July.
When you're ready to step-up to full-frame from an entry-level or midrange camera, the choices can be overwhelming. Find out which models came out on top in our $1200-2000 enthusiast ILC roundup.
Just a guy wearing a VR headset, smashing invisible Goombas in Central Park.
NASA's Mars Reconnaissance Orbiter captured this gorgeous aerial photo of the Martian landscape. And if you look really close, you can actually see the Mars Curiosity rover in the very middle.
The city of Laguna Beach, California has provided some clarification around the kinds of photography permits it offers.
Later this year, a VR180 camera will be Joining Yi's Halo and 360 VR cameras, which will offer stereo 3D capture, yet be as easy to use and compact as a 2D camera.
Caltech researchers have developed an 'optical phased array' chip that uses time delays instead of a lens to focus the incoming light.
Pricing and shipping have finally been revealed for two highly anticipated lenses from Sigma, announced in February.
These macro photos of clouds of paint billowing through clear water might look like high-quality CGI, but they're real photographs. And photographer Alberto Seveso told us how they were made.
Facebook is testing a feature that prevents people from saving, sharing, or even taking a screenshot of your profile picture.
We've reshot the Sony a9 in our studio. The short story: it's sharper! The long story... well you can read it all here.
The collection will be officially launched during the Europeana Transcribathon Campus Berlin 2017 crowdsourcing event which will be held on 22 and 23 June at the Berlin State Library.
Light gives us some insight into the preparations for the launch of the pre-order shipments of its much anticipated L16 multi-lens camera.
OnePlus co-founder Carl Pei has confirmed in a tweet that the second lens on the back of the OnePlus 5 uses a 1.6x optical zoom and that digital zoom is used to reach the claimed 2x zoom factor.
Fujifilm recently unveiled the second in its series of affordable cine lenses, the MK50-135mm T2.9. We got our hands on it for a couple days and took it for a spin.
Leica's first attempt at an M-series digital rangefinder was rough around the edges, but set a pattern for all of the cameras that came after it. In this week's Throwback Thursday article, Barney remembers the M8.
No stranger to extreme situations, legendary climber and filmmaker Jimmy Chin talks to Outside Magazine about his career, and the challenge of filming Alex Honnold's rope-free solo climb of El Capitain.
A company backed by Android co-founder Andy Rubin is attempting to make video conferencing less terrible.
Rangefinder magazine asked five professional portrait and wedding photographers about posting on Instagram; no surprise, they got five different answers.
This captivating stop motion film was created by stripping away one layer of wood at a time. It's hard to look away.
It will enable users to simulate the presence of the sun, moon and Milky Way and see how they interact with an area's topography.