RSS
 

Posts Tagged ‘Updates’

Nikon updates nearly all of its Windows software after anti-virus confusion

03 Dec

A number of Nikon Windows programs have been mistakenly identified as viruses by certain anti-virus programs. While there was no actual threat to the security of the computers the programs were installed on, Nikon has released an update for all of its software products to ensure the programs don’t get incorrectly flagged by anti-virus software going forward.

In the changelog of each of the updates, Nikon says it’s ’Fixed an issue that caused certain anti-virus software to mistakenly identify the “IFGLS001.exe” and “ifasglog.exe” files in the installation destination folder as viruses during or after installation.’

You can download the updates using the following product support pages:

  • Nikon Capture NX-D (Version 1.6.0a)
  • Nikon Camera Control Pro 2 (Version 2.29.1a)
  • Nikon Message Center 2 (Version 2.4.1)
  • Nikon Picture Control Utility 2 (Version 2.4.3a)
  • Nikon ViewNX-i & Capture NX-D (Version 1.19.040)
  • Nikon ViewNX-i (Version 1.4.0a)
  • Nikon Wireless Transmitter Utility (Version 1.9.0a)

Aside from the above change, it doesn’t appear as though anything is different in the programs.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Nikon updates nearly all of its Windows software after anti-virus confusion

Posted in Uncategorized

 

Olympus updates its lens roadmap with three new lenses due out in 2020

28 Nov

Olympus has released an updated version of its M.Zuiko Digital lens roadmap that adds three new lenses to its Micro Four Thirds System lineup.

In addition to its previously-announced lenses, the new roadmap confirms three new lenses: the ED 12-45mm F4.0 Pro standard zoom lens, ED 100-400mm F5.0-6.3 IS super-telephoto zoom lens and ED 150-400mm F4.5 TC1.25x IS PRO lens super-telephoto zoom lens.

A product image from Olympus showing off the impending ED 150-400mm F4.5 TC1.25x IS PRO super-telephoto zoom lens.

Olympus doesn’t elaborate further on any of the new lenses other than saying the detailed specifications will be ‘officially announced in 2020.’ Olympus also says an ‘additional PRO telephoto zoom lens of an unspecified focal length will also join the M.Zuiko family,’ although no timeframe is given for its release.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Olympus updates its lens roadmap with three new lenses due out in 2020

Posted in Uncategorized

 

Peak Design updates its Everyday line with new and improved bags, slings and totes

25 Nov

Peak Design’s Everyday line of camera bags has received a V2 update that includes new materials and features, inspired directly from customer feedback and improvements in material design.

The list of the overhauled bags includes the Everyday Backpack (20L + 30L), Everyday Messenger (15L), Everyday Sling (5L is now 6L + 10L) and the Everyday Tote 15L. In addition to the redesigned originals, Peak Design has also added a number of new bags: Everyday Totepack (15L), Everyday Backpack Zip (15L + 20L) and the Everyday Sling (3L).

All of the bags feature more robust ‘UltraZips,’ improved fabrics that are 100% recycled and Bluesign approved (except for black), improved MagLatch connections and new cord hook external carry straps. The Everyday lineup is also receiving new colorways: Midnight Navy (navy blue) and Bone (tan).

Everyday Backpack

The Everyday Backpack is Peak Design’s most popular design. Although the V2 update isn’t dramatic, the small improvements that have been made do make for an overall improved experience as our hands-on attests to.

The Everyday Backpack V2 features a slightly refined shape that should help it stand upright on its own when filled with gear and has redesigned straps that should better contour your body. It also receives the new UltraZip design, improved MagLatch connections, new internal zippered pockets and the expansion straps found on most of the other V2 bags.

The Everyday Backpack 20L and 30L are available today for $ 259.95 and $ 289.95, respectively.

Everyday Backpack Zip:

The Everyday Backpack Zip is what would happen if you took Peak Design’s Travel Backpack, shrunk it, and added a number of features found in its Everyday line. It comes in 15L and 20L versions and features dual-access zips on either side of the bag for easy gear access. It also has a dedicated top-zip for laptop and tablet access as well as side and top grab handles.

The Everyday Backpack Zip 15L and 20L are available today for $ 189.95 and $ 219.95, respectively.

Everyday Messenger

Similar to the Everyday Backpack, the Everyday Messenger features an almost identical design as its first-generation design but improves upon a number of key features. In addition to the new 100% recycled 400D weather-resistant fabric, the Everyday Messenger features an improved MagLatch design, the new UltraZips and a few small aesthetic changes throughout the bag’s design to clean up the seams found throughout.

Below is a feature rundown of the new Everyday Messenger V2:

The Everyday Messenger 15L is available today for $ 219.95.

Everyday Sling

The Everyday Sling lineup has a number of changes from its V1 iterations. First, the 5L sling has been redesigned to add an extra liter of internal storage, now making it a 6L bag. Peak Design has also added a smaller 3L model that takes the features of the large slings and packs it down into an even smaller form factor.

The 3L and 6L bags can convert to was it packs, while the 10L bag offers a luggage carry pass-through for times when you don’t want it wrapped around your shoulder in the airport.

The Everyday Sling 3L, 6L and 10L are available today for $ 79.95, $ 99.95 and $ 149.95, respectively.

Everyday Tote

Much like the Everyday Messenger, the Everyday Tote isn’t much of a change from the previous version. Aside from a cleaner external design, the more robust UltraZips, the 100% recycled fabric and the external cinch straps the Everyday Tote remains largely unchanged.

The Everyday Tote 15L is available today for $ 149.95.

Everyday Totepack

The Everyday Totepack is more or less an Everyday Tote that’s been made a little larger (20L instead of 15L) and adds a set of straps so it can be worn as a backpack.

The Everyday Totepack features dual zip access points on either side of the bag as well as internal FlexFold dividers and stretchy pockets for organizing your gear as you see fit. It features the 100% recycled 400D weatherproof shell and also has two storable carry cinch straps for those times when you need a little more gear.

The Everyday Totepack 20L is available today for $ 179.95.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Peak Design updates its Everyday line with new and improved bags, slings and totes

Posted in Uncategorized

 

Nikon releases a pair of firmware updates for its Z6, Z7 mirrorless cameras

23 Nov

Nikon has released a pair firmware updates for its Z6 and Z7 full-frame mirrorless camera systems that add new lens support, improves custom settings and address a number of issues.

Firmware version 2.10 (for both the Z6 and Z7) adds support for optical VR with Nikon’s NIKKOR Z DX 16–50mm F3.5–6.3 VR and NIKKOR Z DX 50–250mm F4.5–6.3 VR Z-mount lenses. It also enables lens control rings to adjust ISO sensitivity, a setting that can be found under the ‘Custom Setting Menu.’

Beyond that, the firmware addresses the following issues for both the Z6 and Z7:

  • As of November 14, 2019, the aperture display in the lens info panel for NIKKOR Z 24–70mm f/2.8 S and NIKKOR Z 58mm f/0.95 S Noct lenses shows both the current aperture and adjacent aperture values in modes A and M.
  • When On was selected for Auto ISO sensitivity control and an option that does not include “slow sync” was selected for Flash mode, shutter speed was formerly restricted to values between those chosen for Custom Settings e1 (Flash sync speed) and e2 (Flash shutter speed), but this has now been changed to match the behavior of digital SLR cameras, with the result that the minimum shutter speed now corresponds to the value selected for Auto ISO sensitivity control > Minimum shutter speed.
  • Fixed an issue that in extremely rare cases resulted in the camera failing to correctly record movies with 1920×1080; 120pselected for Frame size/frame rate.
  • Fixed an issue that occasionally resulted in “noise” in the form of white lines at the bottom of the display when the electronic viewfinder was on.
  • Fixed an issue that in rare cases resulted in “noise” in the form of fine horizontal lines appearing throughout the frame in movies and the live view display.
  • Fixed errors in the UTC time zones for the following three cities in the Time zone and date > Time zone display in the SETUP MENU:
    • Caracas: was -4:30, is now -4:00
    • Casablanca: was 00:00, is now +1:00
    • Ankara: was +2:00, is now +3:00
  • Fixed an issue that resulted in an incorrect date of creation being displayed in the Windows 10 “properties” dialogs for movies and time-lapse movies created with the camera, as well as copies created with camera movie editing controls.

You can download the latest firmware updates from Nikon’s Z6 and Z7 support pages.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Nikon releases a pair of firmware updates for its Z6, Z7 mirrorless cameras

Posted in Uncategorized

 

Panasonic releases firmware updates for several MFT, full-frame and compact cameras

20 Nov

Following a series of firmware releases earlier this month, Panasonic has released additional firmware updates for a large number of camera models, including several of its Micro Four Thirds offerings, many compact digital camera models and a couple of full-frame camera models. The updates add a number of new features and improvements ranging from expanded media and wireless transmitter compatibility to improved autofocus performance and ‘enhanced’ video functionality.

The updates were released on Tuesday, November 19, and are available to download from Panasonic now via the links below:

  • DC-GH5: Version 2.6
  • DC-GH5s: Version 1.4
  • DC-G9: Version 2.0
  • DC-GF9/GX800/GX850: Version 1.2
  • DC-GF10/GX880: Version 1.2
  • DMC-GX80/GX85: Version 1.3
  • DC-TZ90/TZ91/TZ92/TZ93/ZS70
  • DC-TZ200/TZ202/TZ220/ZS200/ZS220
  • DMC-LX10/LX15
  • DC-S1: Version 1.3
  • DC-S1R: Version 1.3

Some of the firmware updates are more notable than others. The G9 model, for example, has received support for recording 4K 30p/25p 4:2:2 10-bit video, as well as support for 4K 60p/50p 4:2:2 10-bit HDMI output. The model has also received variable frame rate shooting and HDR video support.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Panasonic releases firmware updates for several MFT, full-frame and compact cameras

Posted in Uncategorized

 

Fujifilm announces upcoming firmware updates for X-T3, X-T30 and other models

14 Nov

Fujifilm has announced plans to release future firmware updates for its X-T3 mirrorless camera. The first update to version 3.10 will arrive in mid-December, according to a release from the company published on Wednesday.

The update scheduled for release in mid-December will bring ‘controls for video recording from Gimbal / Drone via USB communications,’ according to Fujifilm, meaning X-T3 owners will be able to use their camera with a drone and gimbal setup that supports the function. This same functionality will also be released for the X-T30 and other models not specified at this time.

The firmware version 3.20 update is scheduled for release in January 2020 and will bring ‘enhanced autofocus’ with better eye AF frame tracking and face-detection performance, as well as improved autofocus when it comes to foreground subjects. Additionally, versiohn 3.20 will add support for saving up to 9,999 images in each folder on the SD card, a dramatic increase from the current 999 limit.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Fujifilm announces upcoming firmware updates for X-T3, X-T30 and other models

Posted in Uncategorized

 

Adobe updates Lightroom Classic, Mobile and its ChromeOS app at MAX 2019

05 Nov

As part of its MAX 2019 conference, Adobe has announced updates for all of its photography products, including Lightroom on desktop and mobile platforms, Lightroom Classic, and Adobe Camera Raw. A number of changes arrive as part of these updates, including the addition of interactive edits and guided tutorials in Lightroom on Mac and Windows, advanced exporting in Lightroom for Android and ChromeOS and more.

A screenshot from Adobe showing off the new tutorial interface inside Adobe Lightroom.

The majority of the newly announced updates are for Lightroom on desktop. In addition to the aforementioned guided tutorials and interactive edits, this software is being updated with more advanced exporting options including TIF support, GPU acceleration, a new Fill Edges feature for panoramas that is powered by Content-Aware Fill tech, a new migrator tool for moving to Lightroom from Photoshop Elements and support for directly ordering prints from White House Custom Color.

A screenshot from Adobe showing off the interface for the White House Custom Colour inside Adobe Lightroom Classic.

In addition to the desktop software, Adobe is updating Lightroom for ChromeOS, as well as Android and iOS. As mentioned, both Android and ChromeOS have received advanced export control that includes more options and support for batch exporting images in original, JPEG, and TIF formats. These features will arrive on iOS ‘soon,’ according to Adobe.

As well, all three platforms are getting contextual help, which was first added to Lightroom on Mac and Windows earlier this year. The new help menu is accessible by selecting the ‘?’ icon within the Lightroom app on iOS, Android, and ChromeOS.

Lightroom for iOS is finally receiving the Batch Editing feature that launched on desktop last year and on Android and ChromeOS back in August. With this, users can apply changes to multiple image selections simultaneously, cutting down on editing time when using an iPhone or iPad.

Finally, Adobe Camera Raw and Lightroom Classic are both getting the Fill Edges feature for panoramas that is detailed above. Lightroom Classic is also receiving support for multi-batch export and the ability to easily export and share presets and preset groups with others.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Adobe updates Lightroom Classic, Mobile and its ChromeOS app at MAX 2019

Posted in Uncategorized

 

As promised, Canon updates EOS RP firmware to add 24fps recording, additional lens support

01 Nov

As promised earlier this month, Canon has published the latest firmware update for its EOS RP camera that adds support for 24fps video recording.

Firmware version 1.4.0 isn’t a major update, but it fulfills the promise Canon made to add Full HD 23.98p recording to the entry-level full-frame mirrorless camera. The update also adds support for the Canon RF 85mm F1.2L USM DS lens, which is said to arrive before the end of the year.

You can download firmware version 1.4.0 for the Canon EOS RP on Canon’s website for macOS and Windows computers.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on As promised, Canon updates EOS RP firmware to add 24fps recording, additional lens support

Posted in Uncategorized

 

Ilford Photo updates its Multigrade darkroom paper, introduces Ortho Plus film in smaller formats

26 Oct

Black and white film specialist Ilford Photo has announced the development of a new version of its resin-coated Multigrade black and white darkroom printing paper, and said that its existing orthochromatic emulsion will be available in 135 and 120 formats. The company plans to introduce Ilford Multigrade V RC Deluxe as stocks of the current version sell out, and Ilford Ortho Plus should go on sale at the end of November.

The new paper is said to offer deeper blacks and ‘improved’ spacing between the mid-tones, along with a slightly warmer base tint in the paper. Ilford says too that contrast is more consistent across the tonal range which will make printing more predictable.

A sample photo on Ilford Ortho Plus

Ilford has made an orthochromatic emulsion for some time, but it has only been available on sheet formats in the recent past. Cut sheets of 5x4in have been the smallest widely available size, so introducing it for 35mm and 120 medium format cameras will make the film accessible to a much wider audience. Ilford Ortho Plus is designed for copy work, but its color sensitivity also makes it suited to creative uses. Rated at ISO 80 in natural light and ISO 40 in tungsten conditions, the film is only sensitive to green and blue light, and thus darkens anything red in the scene.

It emphases blood stains in forensic work, but can also create interesting effects in portraiture as it darkens freckles and brings out skin texture. The film is very fine-grained and can produce punchy contrast and high resolution in flat copy work. Sensitivity drops off at around 570nm, so it can be processed ‘by eye’ under red safe-light conditions without the risk of fogging. It is suitable for processing with all of Ilford’s main-stream chemicals.

The company has also created a new film processor’s starter kit that includes a Paterson dev tank, two spirals, film clips, measuring cylinders, thermometer and pre-measured chemicals to fit the tank.

For more information see the Ilford Photo website.

Press release

NEW PRODUCTS ANNOUNCED PRODUCTS INCLUDING 5TH GENERATION ILFORD MULTIGRADE RC PAPER & ILFORD ORTHO FILM IN 135 AND 120 FORMATS

We are delighted to announce a number of exciting new products added to our range.

ILFORD MULTIGRADE RC DELUXE

ILFORD announced the world’s first variable contrast MULTIGRADE paper in 1940 and in 2019, as we celebrate our 140th anniversary, we are pleased to announce the 5th generation of our best-selling MULTIGRADE RC Papers.

MULTIGRADE IV RC papers (4th generation) were launched 25 years ago and have been hugely successful becoming the ‘go to’ black & white RC papers for discerning students, amateurs and professional darkroom users all over the world.

“Improving them was not going to be easy but we wanted to bring the manufacturing in line with the latest emulsion making technology used in our other RC and fibre papers and so we have redesigned the emulsion from the ground up. This project has been in development for 8 years and we are extremely pleased and excited with the final product.” Giles Branthwaite, Sales & Marketing Director.

These new papers have a slightly warmer base tint than their predecessor as well as better, deeper blacks, improved mid-grade spacing for easier printing and more consistent contrast throughout the tonal range.

Check your preferred stockist for availability and pricing of the new papers. MULTIGRADE IV RC papers will continue to be available while stocks last.

ILFORD ORTHO PLUS

We are expanding the world’s biggest selection of black & white films with the addition of ILFORD ORTHO PLUS in 135 and 120 formats. As the name suggests this is an orthochromatic black and white film (all other ILFORD and Kentmere films are panchromatic) rated at ISO 80 in natural light and ISO 40 in Tungsten.

“Our Ortho film was designed as a technical, high-resolution copy film for negatives and has been available in sheet format for some time. We know photographers want choice and love to try new films and so we have now coated our Ortho emulsion onto an acetate base for 35mm cassettes and 120 rolls. We believe ORTHO PLUS offers superb photographic potential thanks to its fine grain and sharpness and some of the results we have seen are stunning.” Giles Branthwaite, Sales & Marketing Director.

The blue and green sensitivity of this emulsion enables the film to be handled in deep red safelight conditions making processing and inspection easier. Its lack of red sensitivity also means that reds and oranges are rendered darker than panchromatic films.

ILFORD ORTHO PLUS will be available from early November. Check your preferred stockist for pricing.

ILFORD & Paterson FILM PROCESSING STARTER KIT

Building on the launch of the ILFORD Simplicity range of film processing chemicals in 2018, we have worked in collaboration with Paterson, a specialist manufacturer of some of the world’s most popular darkroom products, to create a starter kit containing all you need to process your first two rolls of film. You just need a light tight space and some film!

“ILFORD Simplicity chemicals are pre-measured for simple, convenient use and while they can easily be used in most tanks their volumes work perfectly with the Paterson 2-reel Universal Film Development Tank as they hold 600ml of diluted chemicals. Working closely with Paterson we have now created a kit that offers film photographers a convenient and affordable way to start processing their own films.” Giles Branthwaite, Sales & Marketing Director.

The kit contains: 1 x ILFORD 35mm Film Cassette Opener, 1 x ILFORD Simplicity Film Starter Pack (containing developer, stop bath, fixer and wetting agent), 1 x Paterson Universal Film Tank, 2 x Paterson Super System 4 Reels, 3 x Paterson 600ml Graduates, 1 x Paterson Thermometer, 1 x Paterson Stirrer, 2 x Paterson set of 2 film clips.

ILFORD & Paterson FILM PROCESSING STARTER KIT will be available from early November. Check your preferred stockist for pricing.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Ilford Photo updates its Multigrade darkroom paper, introduces Ortho Plus film in smaller formats

Posted in Uncategorized

 

These are the most important Google Pixel 4 camera updates

19 Oct

Google yesterday announced the Pixel 4 and Pixel 4 XL, updates to the popular line of Pixel smartphones.

We had the opportunity recently to sit down with Marc Levoy, Distinguished Engineer and Computational Photography Lead at Google, and Isaac Reynolds, Product Manager for Camera on Pixel, to dive deep into the imaging improvements brought to the lineup by the Pixel 4.

Table of contents:

  • More zoom
  • Dual exposure controls / Live HDR+
  • Improved Night Sight
  • DSLR-like bokeh
  • Portrait mode improvements
  • Further improvements
  • Conclusion

Note that we do not yet have access to a production-quality Pixel 4. As such, many of the sample images in this article were provided by Google.

More zoom

The Pixel 4 features a main camera module with a 27mm equivalent F1.7 lens, employing a 12MP 1/2.55″ type CMOS sensor. New is a second ‘zoomed-in’ camera module with a 48mm equivalent, F2.4 lens paired with a slightly smaller 16MP sensor. Both modules are optically stabilized. Google tells us the net result is 1x-3x zoom that is on par with a true 1x-3x optical zoom, and pleasing results all the way out to 4x-6x magnification factors. No doubt the extra resolution of the zoomed-in unit helps with those higher zoom ratios.

Have a look at what the combination of two lenses and super-res zoom gets you with these 1x to 8x full-resolution samples from Google.

$ (document).ready(function() { SampleGalleryV2({“containerId”:”embeddedSampleGallery_3571311037″,”galleryId”:”3571311037″,”isEmbeddedWidget”:true,”selectedImageIndex”:0,”isMobile”:false}) });

Marc emphasized that pinching and zooming to pre-compose your zoomed-in shot is far better than cropping after the fact. I’m speculating here, but I imagine much of this has to do with the ability of super-resolution techniques to generate imagery of higher resolution than any one frame. A 1x super-res zoom image (which you get by shooting 1x Night Sight) still only generates a 12MP image; cropping and upscaling from there is unlikely to get you as good results as feeding crops to the super-res pipeline for it to align and assemble on a higher resolution grid before it outputs a 12MP final image.

We’re told that Google is not using the ‘field-of-view fusion’ technique Huawei uses on its latest phones where, for example, a 3x photo gets its central region from the 5x unit and its peripheries from upscaling (using super-resolution) the 1x capture. But given Google’s choice of lenses, its decision makes sense: from our own testing with the Pixel 3, super-res zoom is more than capable of handling zoom factors between 1x and 1.8x, the latter being the magnification factor of Google’s zoomed-in lens.

Dual exposure controls with ‘Live HDR+’

The results of HDR+, the burst mode multi-frame averaging and tonemapping behind every photograph on Pixel devices, are compelling, retaining details in brights and darks in, usually, a pleasing, believable manner. But it’s computationally intensive to show the end result in the ‘viewfinder’ in real-time as you’re composing. This year, Google has opted to use machine learning to approximate HDR+ results in real-time, leading to a much better viewfinder experience.1 Google calls this ‘Live HDR+’. It’s essentially a WYSIWYG implementation that should give photographers more confidence in the end result, and possibly feel less of a need to adjust the overall exposure manually.

“If we have an intrinsically HDR camera, we should have HDR controls for it” – Marc Levoy

On the other hand, if you do have an approximate live view of the HDR+ result, wouldn’t it be nice if you could adjust it in real-time? That’s exactly what the new ‘dual exposure controls’ allow for. Tap on the screen to bring up two separate exposure sliders. The brightness slider, indicated by a white circle with a sun icon, adjusts the overall exposure, and therefore brightness, of the image. The shadows slider essentially adjusts the tonemap, so you can adjust shadow and midtone visibility and detail to suit your taste.

Default HDR+ result Brightness slider (top left) lowered to darken overall exposure
Shadows slider (top center) lowered to create silhouettes Final result

Dual exposure controls are a clever way to operate an ‘HDR’ camera, as it allows the user to adjust both the overall exposure and the final tonemap in one or two swift steps. Sometimes HDR and tonemapping algorithms can go a bit far (as in this iPhone XS example here), and in such situations photographers will appreciate having some control placed back in their hands.

And while you might think this may be easy to do after-the-fact, we’ve often found it quite difficult to use the simple editing tools on smartphones to push down the shadows we want darkened after tonemapping has already brightened them. There’s a simple reason for that: the ‘shadows’ or ‘blacks’ sliders in photo editing tools may or may not target the same range of tones the tonemapping algorithms did when initially processing the photo.

Improved Night Sight

Google’s Night Sight is widely regarded as an industry benchmark. We consistently talk about its use not just for low light photography, but for all types of photography because of its use of a super-resolution pipeline to yield higher resolution results with less aliasing and moire artifacts. Night Sight is what allowed the Pixel 3 to catch up to 1″-type and four-thirds image quality, both in terms of detail and noise performance in low light, as you can see here (all cameras shot with equivalent focal plane exposure). So how could Google improve on that?

Well, let’s start with the observation that some reviewers of the new iPhone 11 remarked that its night mode had surpassed the Pixel 3’s. While that’s not entirely true, as I covered in my in-depth look at the respective night modes, we have found that at very low light levels the Pixel 3 does fall behind. And it mostly has to do with the limits: handheld exposures per-frame in our shooting with the Pixel 3 were limited to ~1/3s to minimize blur caused by handshake. Meanwhile, the tripod-based mode only allowed shutter speeds up to 1s. Handheld and tripod-based shots were limited to 15 and 6 total frames, respectively, to avoid user fatigue. That meant the longest exposures you could ever take were limited to 5-6s.

Pixel 4 extends the per-frame exposure, when no motion is detected, to at least 16 seconds and up to 15 frames. That’s a total of 4 minutes of exposure. Which is what allows the Pixel 4 to capture the Milky Way:

$ (document).ready(function() { SampleGalleryV2({“containerId”:”embeddedSampleGallery_6422957949″,”galleryId”:”6422957949″,”isEmbeddedWidget”:true,”selectedImageIndex”:0,”isMobile”:false}) });

Remarkable is the lack of user input: just set the phone up against a rock to stabilize it, and press one button. That’s it. It’s important to note you couldn’t get this result with one long exposure, either with the Pixel phone or a dedicated camera, because it would result in star trails. So how does the Pixel 4 get around this limitation?

The same technique that enables high quality imagery from a small sensor: burst photography. First, the camera picks a shutter speed short enough to ensure no star trails. Next, it takes many frames at this shutter speed and aligns them. Since alignment is tile-based, it can handle the moving stars due to the rotation of the sky just as the standard HDR+ algorithm handles motion in scenes. Normally, such alignment is very tricky for photographers shooting night skies with non-celestial, static objects in the frame, since aligning the stars would cause misalignment in the foreground static objects, and vice versa.

Improved Night Sight will not only benefit starry skyscapes, but all types of photography requiring long exposures

But Google’s robust tile-based merge can handle displacement of objects from frame to frame of up to ~8% in the frame2. Think of it as tile-based alignment where each frame is broken up into roughly 12,000 tiles, with each tile individually aligned to the base frame. That’s why the Pixel 4 has no trouble treating stars in the sky differently from static foreground objects.

Another issue with such long total exposures is hot pixels. These pixels can become ‘stuck’ at high luminance values as exposure times increase. The new Night Sight uses clever algorithms to emulate hot pixel suppression, to ensure you don’t have bright pixels scattered throughout your dark sky shot.

DSLR-like bokeh

This is potentially a big deal, and perhaps underplayed, but the Google Pixel 4 will render bokeh, particularly out-of-focus highlights, closer to what we’d expect from traditional cameras and optics. Until now, while Pixel phones did render proper disc-shaped blur for out of focus areas as real lenses do (as opposed to a simple Gaussian blur), blurred backgrounds simply didn’t have the impact they tend to have with traditional cameras, where out-of-focus highlights pop out of the image in gorgeous, bright, disc-shaped circles as they do in these comparative iPhone 11 examples here and also here.

The new bokeh rendition on the Pixel 4 takes things a step closer to traditional optics, while avoiding the ‘cheap’ technique some of its competitors use where bright circular discs are simply ‘stamped’ in to the image (compare the inconsistently ‘stamped’ bokeh balls in this Samsung S10+ image here next to the un-stamped, more accurate Pixel 3 image here). Have a look below at the improvements over the Pixel 3; internal comparisons graciously provided to me via Google.

Daytime bokeh

Daytime bokeh

Nighttime bokeh

Nighttime bokeh

The impactful, bright, disc-shaped bokeh of out-of-focus highlights are due to the processing of the blur at a Raw level, where linearity ensures that Google’s algorithms know just how bright those out-of-focus highlights are relative to their surroundings.

Previously, applying the blur to 8-bit tonemapped images resulted in less pronounced out-of-focus highlights, since HDR tonemapping usually compresses the difference in luminosity between these bright highlights and other tones in the scene. That meant that out-of-focus ‘bokeh balls’ weren’t as bright or separated from the rest of the scene as they would be with traditional cameras. But Google’s new approach of applying the blur at the Raw stage allows it to more realistically approximate what happens optically with conventional optics.

One thing I wonder about: if the blur is applied at the Raw stage, will we get Raw portrait mode images in a software update down-the-line?

Portrait mode improvements

Portrait mode has been improved in other ways apart from simply better bokeh, as outlined above. But before we begin I want to clarify something up front: the term ‘fake bokeh’ as our readers and many reviewers like to call blur modes on recent phones is not accurate. The best computational imaging devices, from smartphones to Lytro cameras (remember them?), can actually simulate blur true to what you’d expect from traditional optical devices. Just look at the gradual blur in this Pixel 2 shot here. The Pixel phones (and iPhones as well as other phones) generate actual depth maps, gradually blurring objects from near to far. This isn’t a simple case of ‘if area detected as background, add blurriness’.

The Google Pixel 3 generated a depth map from its split photodiodes with a ~1mm stereo disparity, and augmented it using machine learning. Google trained a neural network using depth maps generated by its dual pixel array (stereo disparity only) as input, and ‘ground truth’ results generated by a ‘franken-rig’ that used 5 Pixel cameras to create more accurate depth maps than simple split pixels, or even two cameras, could. That allowed Google’s Portrait mode to understand depth cues from things like defocus cues (out-of-focus objects are probably further away than in-focus ones) and semantic cues (smaller objects are probably further away than larger ones).

Deriving stereo disparity from two perpendicular baselines affords the Pixel 4 much more accurate depth maps

The Pixel 4’s additional zoomed-in lens now gives Google more stereo data to work with, and Google has been clever in its arrangement: if you’re holding the phone upright, the two lenses give you horizontal (left-right) stereo disparity, while the split pixels on the main camera sensor give you vertical (up-down) stereo disparity. Having stereo data along two perpendicular axes avoids artifacts related to the ‘aperture problem’, where detail along the axis of stereo disparity essentially has no measured disparity.

Try this: look at a horizontal object in front of you and blink to switch between your left and right eye. The object doesn’t look very different as you switch eyes, does it? Now hold out your index finger, pointing up, in front of you, and do the same experiment. You’ll see your finger moving dramatically left and right as you switch eyes.

Deriving stereo disparity from two perpendicular baselines affords the Pixel 4 much more accurate depth maps, with the dual cameras providing disparity information that the split pixels might miss, and vice versa. In the example below, provided by Google, the Pixel 4 result is far more believable than the Pixel 3 result, which has parts of the upper and lower green stem, and the horizontally-oriented green leaf near bottom right, accidentally blurred despite falling within the plane of focus.

(dual baseline)

(single baseline)

The combination of two baselines, one short (split pixels) and one significantly longer (the two lenses) also has other benefits. The longer stereo baselines of dual camera setups can run into the problem of occlusion: since the two perspectives are considerably different, one lens may see a background object that to the other lens is hidden behind a foreground object. The shorter 1mm disparity of the dual pixel sensor means its less prone to errors due to occlusion.

On the other hand, the short disparity of the split pixels means that further away objects that are not quite at infinity appear the same to ‘left-looking’ and ‘right-looking’ (or up/down) photodiodes. The longer baseline of the dual cameras means that stereo disparity can be calculated for these further away objects, which allows the Pixel 4’s portrait mode to better deal with distant subjects, or groups of people shot from further back, as you can see below.

There’s yet another benefit of the two separate methods for calculating stereo disparity: macro photography. If you’ve shot portrait mode on telephoto units of other smartphones, you’ve probably run into error messages like ‘Move farther away’. That’s because these telephoto lenses tend to have a minimum focus distance of ~20cm. Meanwhile, the minimum focus distance of the main camera on the Pixel 4 is only 10cm. That means that for close-up photography, the Pixel 4 can simply use its split pixels and learning-based approach to blur backgrounds.3

One thing we’ll be curious to test is if the additional burden of taking two images with the dual camera setup will lead to any latency. The iPhone 11, for example, has considerable shutter lag in portrait mode.

Google continues to keep a range of planes in perfect focus, which can sometimes lead to odd results where multiple people in a scene remain focused despite being at different depths. However, this approach avoids prematurely blurring parts of people that shouldn’t be blurred, a common problem with iPhones.

Oddly, portrait mode is unavailable with the zoomed-in lens, instead opting to use the same 1.5x crop from the main camera that the Pixel 3 used. This means images will have less detail compared to some competitors, especially since the super-res zoom pipeline is still not used in portrait mode. It also means you don’t get the versatility of both wide-angle and telephoto portrait shots. And if there’s one thing you probably know about me, it’s that I love my wide angle portraits!

Pixel 4’s portrait mode continues to use a 1.5x crop from the main camera. This means that, like the Pixel 3, it will have considerably less detail than portrait modes from competitors like the iPhone 11 Pro that use the full-resolution image from wide or tele modules. Click to view at 100%

Further improvements

There are a few more updates to note.

Learning-based AWB

The learning-based white balance that debuted in Night Sight is now the default auto white balance (AWB) algorithm in all camera modes on the Pixel 4. What is learning-based white balance? Google trained its traditional AWB algorithm to discriminate between poorly, and properly, white balanced images. The company did this by hand-correcting images captured using the traditional AWB algorithm, and then using these corrected images to train the algorithm to suggest appropriate color shifts to achieve a more neutral output.

Google tells us that the latest iteration of the algorithm is improved in a number of ways. A larger training data set has been used to yield better results in low light and adversarial lighting conditions. The new AWB algorithm is better at recognizing specific, common illuminants and adjusting for them, and also yields better results under artificial lights of one dominant color. We’ve been impressed with white balance results in Night Sight on the Pixel 3, and are glad to see it ported over to all camera modes. See below how Google’s learning-based AWB (top left) preserves both blue and red/orange tones in the sky compared to its traditional AWB (top right), and how much better it is at separating complex sunset colors (bottom left) compared to the iPhone XS (bottom right).

Learning-based AWB (Pixel 3 Night Sight) Traditional AWB (Pixel 3)
Learning-based AWB (Pixel 3 Night Sight) iPhone XS HDR result

New face detector

A new face detection algorithm based solely on machine learning is now used to detect, focus, and expose for faces in the scene. The new face detector is more robust at identifying faces in challenging lighting conditions. This should help the Pixel 4 better focus on and expose for, for example, strongly backlit faces. The Pixel 3 would often prioritize exposure for highlights and underexpose faces in backlit conditions.

Though tonemapping would brighten the face properly in post-processing, the shorter exposure would mean more noise in shadows and midtones, which after noise reduction could lead to smeared, blurry results. In the example below the Pixel 3 used an exposure time of 1/300s while the iPhone 11 yielded more detailed results due to its use of an exposure more appropriate for the subject (1/60s).

Along with the new face detector, the Pixel 4 will (finally) indicate the face it’s focusing on in the ‘viewfinder’ as you compose. In the past, Pixel phones would simply show a circle in the center of the screen every time it refocused, which was a very confusing experience that left users wondering whether the camera was in fact focusing on a face in the scene, or simply on the center. Indicating the face its focusing on should allow Pixel 4 users to worry less, and feel less of a need to tap on a face in the scene if the camera’s already indicating it’s focusing on it.

On previous Pixel phones, a circle focus indicator would pop up in the center when the camera refocused, leading to confusion. Is the camera focusing on the face, or the outstretched hand? On the Huawei P20, the camera indicates when it’s tracking a face. The Pixel 4 will have a similar visual indicator.

Semantic segmentation

This isn’t new, but in his keynote Marc mentioned ‘semantic segmentation’ which, like the iPhone, allows image processing to treat different portions of the scene differently. It’s been around for years in fact, allowing Pixel phones to brighten faces (‘synthetic fill flash’), or to better separate foregrounds and backgrounds in Portrait mode shots. I’d personally point out that Google takes a more conservative approach in its implementation: faces aren’t brightened or treated differently as much as they tend to be with the iPhone 11. The end result is a matter of personal taste.

Conclusion

The questions on the minds of many of our readers will undoubtedly be: (1) what is the best smartphone for photography I can buy, and (2) when should I consider using such a device as opposed to my dedicated camera?

We have much testing to do and many side-by-sides to come. But from our tests thus far and our recent iPhone 11 vs. Pixel 3 Night Sight article, one thing is clear: in most situations the Pixel cameras are capable of a level of image quality unsurpassed by any other smartphone when you compare images at the pixel (no pun intended) level.

But other devices are catching up, or exceeding Pixel phone capabilities. Huawei’s field-of-view fusion offers compelling image quality across multiple zoom ratios thanks to its fusion of image data from multiple lenses. iPhones offer a wide-angle portrait mode far more suited for the types of photography casual users engage in, with better image quality to boot than Pixel’s (cropped) Portrait mode.

The Pixel 4 takes an already great camera and refines it to achieve results closer to, and in some cases surpassing, traditional cameras and optics

Overall though, Google Pixel phones deliver some of the best image quality we’ve seen from a mobile device. No other phone can compete with its Raw results, since Raws are a result of a burst of images stacked using Google’s robust align-and-merge algorithm. Night Sight is now improved to allow for superior results with static scenes demanding long exposures. And Portrait mode is vastly improved thanks to dual baselines and machine learning, with fewer depth map errors and better ability to ‘cut around’ complex objects like pet fur or loose hair strands. And pleasing out-of-focus highlights thanks to ‘DSLR-like bokeh’. AWB is improved, and a new learning-based face detector should improve focus and exposure of faces under challenging lighting.

It’s not going to replace your dedicated camera in all situations, but in many it might. The Pixel 4 takes an already great camera in the Pixel 3, and refines it further to achieve results closer to, and in some cases surpassing, traditional cameras and optics. Stay tuned for more thorough tests once we get a unit in our hands.

Finally, have a watch of Marc Levoy's Keynote presentation yesterday below. And if you haven’t already, watch his lectures on digital photography or visit his course website from the digital photography class he taught while at Stanford. There’s a wealth of information on digital imaging in those talks, and Marc has a knack for distilling complex topics into elegantly simple terms.


Footnotes:

1 The Pixel 3’s dim display combined with the dark shadows of a non-HDR preview often made the experience of shooting high contrast scenes outdoors lackluster, sometimes even making it difficult to compose. Live HDR+ should dramatically improve the experience, though the display remains relatively dim compared to the iPhone 11 Pro.

2 The original paper on HDR+ by Hasinoff and Levoy claims HDR+ can handle displacements of up to 169 pixels within a single raw color channel image. For a 12MP 4:3 Bayer sensor, that’s 169 pixels of a 2000 pixel wide (3MP) image, which amounts to ~8.5%. Furthermore, tile-based alignment is performed using as small as 16×16 pixel blocks of that single raw channel image. That amounts to ~12,000 effective tiles that can be individually aligned.

3 The iPhone 11’s wide angle portrait mode also allows you to get closer to subjects, since its ultra-wide and wide cameras can focus on nearer subjects than its telephoto lens.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on These are the most important Google Pixel 4 camera updates

Posted in Uncategorized