Apple Research Uncovers Possibilities of AI-Enhanced ISP to Improve Low-Light Photography on iPhones
# Apple’s DarkDiff: Transforming Low-Light Photography with AI
Apple’s researchers have made remarkable progress in improving extremely dark images through the creation of an AI model named DarkDiff. This groundbreaking technique incorporates a diffusion-based image model directly within the camera’s image processing pipeline, allowing for the retrieval of details from raw sensor data that usually gets lost in low-light scenarios.
## The Challenge of Extreme Low-Light Photos
Taking photos in very dark settings often results in grainy images filled with digital noise. This happens when the image sensor does not capture enough light, leading to a loss of detail. To mitigate this, companies like Apple have turned to image processing algorithms. However, these algorithms have been criticized for producing overly smooth images that lack intricate detail, resulting in a painted effect where the original content becomes unrecognizable.
## Introducing DarkDiff
To tackle the difficulties of low-light photography, researchers from Apple and Purdue University unveiled DarkDiff, as outlined in their study titled “DarkDiff: Advancing Low-Light Raw Enhancement by Retasking Diffusion Models for Camera ISP.” The researchers emphasize that achieving high-quality photography in extreme low-light conditions remains a challenge, but improvements in computing hardware are facilitating smarter enhancement of noisy raw images.
DarkDiff differs from conventional post-processing techniques by repurposing Stable Diffusion, an open-source model trained on millions of images. This model is engineered to comprehend which details should appear in dark areas of photos based on their overall context and is integrated into the image signal processing (ISP) pipeline of the camera.
The DarkDiff framework calculates attention over localized image patches, which helps in maintaining local structures and minimizing hallucinations—artifacts where the AI inaccurately modifies image content. The ISP first processes the raw sensor data, executing crucial steps like white balance and demosaicing, before DarkDiff works on the linear RGB image to denoise it and create the final sRGB image.
DarkDiff utilizes a technique known as classifier-free guidance, which dictates how closely the model should conform to the input image in contrast to its learned visual priors. Modifying this guidance enables a balance between generating smoother patterns and sharper textures, though it also heightens the risk of introducing unwanted artifacts.
## Evaluating DarkDiff
The researchers assessed DarkDiff using actual photos taken in extremely low-light situations with cameras such as the Sony A7SII. They compared the outcomes against other raw enhancement models and diffusion-based benchmarks, including ExposureDiff. The test images were captured at night with exposure times as short as 0.033 seconds, and DarkDiff’s improvements were reviewed against reference photos taken with significantly prolonged exposure times on a tripod.
## DarkDiff’s Challenges
Despite its promising capabilities, the researchers recognize that DarkDiff’s AI-driven processing is significantly slower than traditional methods, possibly requiring cloud processing to manage its high computational needs. Running the model locally on a smartphone could rapidly drain battery life. Furthermore, there are constraints regarding non-English text recognition in low-light environments.
While the study does not suggest that DarkDiff will be implemented in iPhones in the immediate future, it emphasizes Apple’s ongoing dedication to advancing computational photography—a vital area of focus in the smartphone market as consumers demand camera functionalities that exceed the physical limits of their devices.
For those interested in an in-depth exploration, the complete study and further comparisons between DarkDiff and other denoising methods can be found
Read More