Rather than concentrating on any one hardware aspect of iPhone photography, Apple’s engineers and managers aim to control how the company manages every step of taking a photo.
With the launch of the iPhone 12 Pro Max, Apple has introduced the largest camera sensor it has ever put in an iPhone. Yet rather than being there to “brag about,” Apple says that it is part of a philosophy that sees camera designers working across every possible aspect from hardware to software.
Speaking to photography site PetaPixel, Francesca Sweet, product line manager for the iPhone, and Jon McCormack, vice president of camera software engineering, emphasized that they work across the whole design in order to simplify taking photos.
“As photographers, we tend to have to think a lot about things like ISO, subject motion, et cetera,” Job McCormack said. “And Apple wants to take that away to allow people to stay in the moment, take a great photo, and get back to what they’re doing.”
“It’s not as meaningful to us anymore to talk about one particular speed and feed of an image, or camera system,” he continued. “We think about what the goal is, and the goal is not to have a bigger sensor that we can brag about.”
“The goal is to ask how we can take more beautiful photos in more conditions that people are in,” he said. “It was this thinking that brought about Deep Fusion, Night Mode, and temporal image signal processing.”
Apple’s overall aim, both McCormack and Sweet say, is to automatically “replicate as much as we can… what the photographer will [typically] do in post.” So with Machine Learning, Apple’s camera system breaks down an image into elements that it can then process.
“The background, foreground, eyes, lips, hair, skin, clothing, skies,” lists McCormack. “We process all these independently like you would in [Adobe] Lightroom with a bunch of local adjustments. We adjust everything from exposure, contrast, and saturation, and combine them all together.”
This isn’t to deny the advantages of a bigger sensor, according to Sweet. “The new wide camera [of the iPhone 12 Pro Max], improved image fusion algorithms, make for lower noise and better detail.”
“With the Pro Max we can extend that even further because the bigger sensor allows us to capture more light in less time, which makes for better motion freezing at night,” she continued.
Apple’s iPhone 12 range brings camera improvements across the board
Nonetheless, both Sweet and McCormack believe that it is vital how Apple designs and controls every element from lens to software.
“We don’t tend to think of a single axis like ‘if we go and do this kind of thing to hardware’ then a magical thing will happen,” said McCormack. “Since we design everything from the lens to the GPU and CPU, we actually get to have many more places that we can do innovation.”