Saturday 2 November 2019

What Is the Deep Fusion Camera on the iPhone 11?

iPhone 11 Pro Max Cameras
Apple

Apple’s Deep Fusion Camera technology is now available thanks to iOS 13.2. If you have an iPhone 11 or iPhone 11 Pro, you can use this new image-processing tech to take better photos. Here’s how it works.

What Is Deep Fusion?

Smartphones aren’t complete replacements to professional cameras just yet, but Apple makes the iPhone a better camera every year.

This feature is available on the iPhone 11, iPhone 11 Pro, and iPhone 11 Pro Max. These phones were released with Apple’s iOS 13. They came with several significant improvements to their camera setup, including improved sensors, an ultrawide-angle lens, a night mode, and slow-motion selfies. However, one improvement that didn’t come out of the box with their newest flagships is the Deep Fusion Camera, released with the iOS 13.2 update on October 28, 2019.

Apple’s Phil Schiller described it as “computational photography mad science.” While many smartphones are making great strides towards improving image quality in very dark environments with Night Mode and very bright environments with HDR, most of the photos we take fall somewhere in between. The Deep Fusion Camera is supposed to reduce noise and significantly improve detail for photos taken in medium to low-light conditions, mainly indoor shots.

To demonstrate, Apple used several samples of people wearing sweaters—an item of clothing that frequently loses detail in photos. The sweaters and other items in the shots taken with the Deep Fusion Camera are more detailed and retain their natural texture.

RELATED: The Best New Features in iOS 13, Available Now

How Does It Work?

Woman Wearing Knitted Sweater - iPhone Deep Fusion
Apple via The Verge

According to Apple, the new mode uses the iPhone 11’s new A13 Bionic chip to do “pixel-by-pixel processing of photos, optimizing for texture, details, and noise in every part of the photo.” In essence, it works similarly to the iPhone camera’s Smart HDR, which takes several shots at varying exposures and combines them to maximize the clarity in the finished image. Where they differ is in the amount of information that needs to be processed.

What Deep Fusion is doing in the background is quite complicated. When the user presses the shutter button in medium light, the camera immediately takes nine pictures: four short images, four secondary images, and one long exposure photo. It fuses the long-exposure with the best among the short images. Then, the processor goes pixel by pixel and selects the best elements from both to create the most detailed photo possible. All of that takes place in one second.

When you initially snap a photo, it immediately starts post-processing the image in your album. So by the time you open your camera roll to take a look at it, it will already have the effect implemented. This is made possible by the A13 Bionic chip, which is the strongest processor ever put into a commercial smartphone.

How to Get Deep Fusion on Your iPhone

Read the remaining 14 paragraphs



from How-To Geek https://ift.tt/2PC3jis

No comments:

Post a Comment