• Google has blogged about its recent improvements in AI and photography — specifically regarding portrait mode on the Pixel 3.
  • The post discusses how Google has improved the way its neural networks measure depth.
  • The result is an improved bokeh effect in its portrait mode shots.


Google has detailed one of the major photography accomplishments it achieved on the Pixel 3 on its AI blog. In the post published yesterday, Google discussed how it improved portrait mode between the Pixel 2 and the Pixel 3.

Portrait mode is a popular smartphone photography mode that blurs the background of a scene while maintaining focus on the foreground subject (what’s sometimes called the bokeh effect). The Pixel 3 and the Google Camera app take advantage of advances in neural networks, machine learning, and GPU hardware to help make this effect even better.

In Portrait Mode on the Pixel 2, the camera would capture two versions of a scene at slightly different angles. In these images, the foreground figure, a person in most portrait images, would appear to shift to a smaller degree than the background images (an effect known as parallax). This discrepancy was used as the basis for interpreting the depth of an image, and thus which areas to blur out.

A Google Portrait Mode parallax scrolling example. Google Blog

This provided strong results on the Pixel 2, but it wasn’t perfect. The two versions of the scene provided only a very small amount of information about the depth, so problems could occur. Most commonly, the Pixel 2 (and many others like it) would fail to accurately separate foreground from background.

With the Google Pixel 3’s camera, Google included more depth cues to inform this blur effect for greater accuracy. As well as parallax, Google used sharpness as a depth indicator — more distant objects are less sharp than closer objects — and real-world object identification. For example, the camera could recognize a person’s face in a scene, and work out how near or far it was based on its number of pixels relative to objects around it. Clever.

Google then trained its neural network with the help of the new variables to give it a better understanding — or rather, estimation — of depth in an image.

Google Pixel 3 portrait mode bokeh skull

The Pixel’s portrait mode doesn’t just require a human subject.

What does it all mean?

The result is better-looking portrait mode shots when using the Pixel 3 compared to previous Pixel (and ostensibly many other Android phones) cameras thanks to more accurate background blurring. And, yes, this should mean less hair being lost to background blur.

Editor's Pick

There’s an interesting implication of all this which relates to chips too. A lot of power is required to process the necessary data to create these photographs once they’re snapped (they’re based on full resolution, multi-megapixel PDAF images); the Pixel 3 handles this pretty well thanks to its TensorFlow Lite and GPU combination.

In the future, though, better processing efficiency and dedicated neural chips will widen the possibilities not only for how quickly these shots will be delivered, but for what enhancements developers even choose to integrate.

To find out more about the Pixel 3 camera, hit the link, and give us your thoughts on it in the comments.

This comes from our sister site, dgit Daily. Over there we offer all the latest updates from accross the techiverse.

Visit dgit Daily

Finally, a tech subscription worth reading.

Sign up for daily digests of the tech content most relevant to you.
By signing up, you agree to our Privacy Policy and European users agree to the data transfer policy.