RSS
 

Posts Tagged ‘algorithm’

New DAIN algorithm generates near-perfect slow-motion videos from ordinary footage

09 Sep

Researchers with Google, UC Merced and Shanghai Jiao Tong University have detailed the development of DAIN, a depth-aware video frame interpolation algorithm that can seamlessly generate slow-motion videos from existing content without introducing excessive noise and unwanted artifacts. The algorithm has been demonstrated in a number of videos, including historical footage boosted to 4K/60fps.

Rapidly advancing technologies have paved the way for high-resolution displays and videos; the result is a mass of lower-resolution content made for older display and video technologies that look increasingly poor on modern hardware. Remastering this content to a higher resolution and frame rate will improve the viewing experience, but would typically be a costly undertaking reserved only for the most popular media.

Artificial intelligence is a promising solution for updating older video content as evidenced by the growing number of fan-remastered movies and TV shows. Key to these efforts are algorithms trained to upscale and, when necessary, repair the individual frames of videos, which are recompiled into a higher-resolution ‘remaster.’

The newly detailed DAIN algorithm is different — rather than upscaling and repairing the individual frames in a video, this AI tool works by generating new frames and slotting them between the original frames, increasing the video’s FPS for smoother and, depending on how many frames are generated, slower-motion content.

This is a process called motion (video frame) interpolation, and it typically causes a drop in quality by adding unwanted noise and artifacts to the final videos. The DAIN algorithm presents a solution to this problem, offering motion interpolation to boost frames-per-second up to 480fps without introducing any readily noticeable artifacts.

The resulting content is high-quality and nearly visually identical to the source footage, but with the added smoothness that comes with increasing the frames-per-second to 60fps. In addition, DAIN has been demonstrated as capable of transforming ordinary 30/60fps footage into smooth slow-motion videos without choppiness or decreased quality.

According to the researchers, DAIN is ‘compact, efficient, and fully differentiable,’ offering results that perform ‘favorably against state-of-the-art frame interpolation methods on a wide variety of datasets.’ The technology has many potential uses, including recovering lost frames, improving content to be more visually appealing for viewers, generating slow-motion from regular footage and more.

Such technology is arguably necessary for preserving aging media in a useful way, making it possible for new generations of people to experience historical footage, old TV shows and movies, home videos and similar content using modern high-resolution displays. As well, the technology could be useful for content creators of all sorts, enabling them to salvage the footage they already have, improve the quality of old clips for use in documentaries and similar things.

The researchers explain on their project website:

Starting from the birth of photographing in the 18-th centuries, videos became important media to keep vivid memories of their age being captured. And it’s shown in varying forms including movies, animations, and vlogs. However, due to the limit of video technologies including sensor density, storage and compression, quite a lot of video contents in the past centuries remain at low quality.

Among those important metrics for video quality, the most important one is the temporal resolution measured in frame-per-second or fps for short. Higher-frame-rate videos bring about more immersive visual experience to users so that the reality of the captured content is perceived. Therefore, the demand to improve the low-frame-rate videos, particularly the 12fps old films, 5~12fps animations, pixel-arts and stop motions, 25~30 fps movies, 30fps video games, becomes more and more urgent.

The public can view more examples of videos updated using the DAIN algorithm by checking out the related collection playlist on YouTube. As well, the full study is available in PDF form on the Arxiv website here.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on New DAIN algorithm generates near-perfect slow-motion videos from ordinary footage

Posted in Uncategorized

 

News: Could This Sea-Thru Algorithm Be the Future of Underwater Photography?

18 Nov

The post News: Could This Sea-Thru Algorithm Be the Future of Underwater Photography? appeared first on Digital Photography School. It was authored by Jaymes Dempsey.

Image: Photo by James Thornton via Unsplash.

Photo by James Thornton via Unsplash.

Underwater photographers may soon have a way of “removing water” from their photos, based on research done by Derya Akkaynak and Tali Treibitz at the University of Haifa (read report here).

As explained by Akkaynak and Treibitz, “An underwater photo is the equivalent of one taken in air, but covered in thick, colored fog.” And while the precise effects of water on images is somewhat technical, it doesn’t take much to recognize that water degrades images, causing a loss of both clarity and accurate color.

Enter Akkaynak’s Sea-thru algorithm, which is designed to remove color casts and other optical problems created by water. In other words, it can be applied to an underwater photo, one that’s blue (with inaccurate colors), and turn it into something that looks like it was taken on land.

For examples, check out the images in the Scientific American video:

??

How it works

But how was the algorithm actually developed, and how does it work?

Basically, Akkaynak took a series of underwater photos, making sure to place her color chart into the scene for an accurate reference. She ultimately compiled over a thousand images in several environments. From the reference images, Akkaynak and Treibitz developed a model that takes into account the unique ways that light interacts with water in order to correct underwater images for color and light.

Now, Akkaynak and Treibitz had academic purposes in mind when they conducted this research. The algorithm, as presented in the original research paper, is meant to “help boost underwater research at a time when our oceans are increasing stress from pollution, overfishing, and climate change,” by giving researchers better access to visual data from underwater cameras.

But it’s easy to see how the Sea-thru algorithm could be relevant to underwater photographers everywhere. If Sea-thru can make photos become more accurate and (often) more vivid and colorful, might underwater photographers like to use it on their own images?

On the other hand, there’s the question of whether the best underwater photos convey an authentic sense of the (underwater) environment. Without the blue tones of water and the haze that water provides, photos may lose the sense of wonder that comes from doing work under the sea.

So let me ask you: Do you prefer underwater images where the water is much less apparent? Or do you like more authentic underwater photos, color cast and all? Would you be interested in the Sea-thru software?

Share your thoughts in the comments!

The post News: Could This Sea-Thru Algorithm Be the Future of Underwater Photography? appeared first on Digital Photography School. It was authored by Jaymes Dempsey.


Digital Photography School

 
Comments Off on News: Could This Sea-Thru Algorithm Be the Future of Underwater Photography?

Posted in Photography

 

Video: Google’s Super Resolution algorithm explained in three minutes

30 May

Space constraints in the thin bodies of modern smartphones mean camera engineers are limited in terms of the size of image sensors they can use in their designs. Manufacturers have therefore been pushing computational imaging methods in order to improve the quality of their devices’ image output.

Google’s Super Resolution algorithm is one such method. It involves shooting a burst of raw photos every time the shutter is pressed and takes advantage of the user’s natural hand-shake, even if it is ever so slight. The pixel-level differences between each of the frames in the burst can be used to merge several images of the burst into an output file with optimized detail at each pixel location.

An illustration that shows how multiple frames are aligned to create the final image.

Google uses the Super Resolution in the Night Sight feature and Super-Res zoom of the Pixel 3 devices and has previously published an in-depth article about it on its blog . Our own Rishi Sanyal has also had a close look at the technology and the features it has been implemented in.

A visual representation of the steps used to create the final image from a burst of Raw input images.

Now Google has published the above video that provides a great overview of the the technology in just over three minutes.

‘This approach, which includes no explicit demosaicing step, serves to both increase image resolution and boost signal to noise ratio,’ write the Google researchers in the paper the video is based on. ‘Our algorithm is robust to challenging scene conditions: local motion, occlusion, or scene changes. It runs at 100 milliseconds per 12-megapixel RAW input burst frame on mass-produced mobile phones.’

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Video: Google’s Super Resolution algorithm explained in three minutes

Posted in Uncategorized

 

This NVIDIA algorithm copies the artistic style of one photo onto another

26 Mar

Struggling with stylistic consistency, or wanting to transpose the style of your best picture onto the rest of your Instagram feed? Thanks to a group of scientists at Cornell University, you can now do just that with surprisingly accurate and realistic results.

The team created an algorithm for graphics card company NVIDIA that lifts the stylistic characteristics of one picture and drops them onto a completely different image with startling precision. The algorithm is called FastPhotoStyle, and it’s capable of transferring the coloration, drama and atmosphere of one picture and making an entirely different frame look as though it was taken at the same time even if the subject matter is totally unrelated.

According to the developers, the goal of photorealistic image style transfer is:

…to change the style of a photo to resemble that of another one. For a faithful stylization, the content in the output photo should remain the same, while the style of the output photo should resemble the one of the reference photo. Furthermore, the output photo should look like a real photo captured by a camera.

There are programs already invented to do this, but the inventors of this algorithm claim that what already exists is slow, and doesn’t produce realistic results anyhow.

FastPhotoStyle is different, they say, because it uses a smoothing process after the initial whitening and Coloring Transfer step—or PhotoWCT step. This smoothing step tries to ensure that neighboring pixels receive similar styling and, by using what they call Matting Affinity, individual areas of the image can be subjected to slightly different treatment. This is what helps the algorithm produce such realistic looking results.

Another major difference is that this program reportedly operates as much as 60x faster than existing algorithms.

The code can be downloaded from NVIDIA’s GitHub for anyone to use under Creative Commons license (BY-NC-SA 4.0), and a user manual download is included on the page. If you’re brave, you can read the full technical paper as well.

Technical Paper Abstract:

A Closed-Form Solution to Photorealistic Image Stylization

Photorealistic image style transfer algorithms aim at stylizing a content photo using the style of a reference photo with the constraint that the stylized photo should remains photorealistic.

While several methods exist for this task, they tend to generate spatially inconsistent stylizations with noticeable artifacts. In addition, these methods are computationally expensive, requiring several minutes to stylize a VGA photo. In this paper, we present a novel algorithm to address the limitations.

The proposed algorithm consists of a stylization step and a smoothing step. While the stylization step transfers the style of the reference photo to the content photo, the smoothing step encourages spatially consistent stylizations. Unlike existing algorithms that require iterative optimization, both steps in our algorithm have closed-form solutions.

Experimental results show that the stylized photos generated by our algorithm are twice more preferred by human subjects in average. Moreover, our method runs 60 times faster than the state-of-the-art approach.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on This NVIDIA algorithm copies the artistic style of one photo onto another

Posted in Uncategorized

 

US Megaregions: Algorithm Redefines Boundaries of Metropolitan Areas

07 Dec

[ By WebUrbanist in Culture & History & Travel. ]

real-america

A new geographical study of the United States reveals the functional boundaries of megapolises around the country, defining them by usage rather than arbitrary political borders. Unlike gerrymandered districts or state lines, these sprawling areas are rooted in deep data analytics versus historical accident.

america-borders

Historical geographer Garrett Dash Nelson teamed up with urban analyst Alasdair Rae to publish a paper using commuting information and computational algorithms. Studying over 4,000,000 commutes, they traced interconnections between economically connected points and reported the results in An Economic Geography of the United States: From Commutes to Megaregions.

border-edges

Taking it a step further, the authors also devised names for various megaregions extrapolated from the data – while semi-subjective, they start to give a sense of the real shape of metropolitan zones (and reveal areas where few residents and vast distances make it hard to define or confine regions).

connections-raw

Some cities at the heart of various sub-regions are not surprising — San Francisco and Los Angeles were givens — but others may be new to some people, like Fresno, California. Many cities trace influence across state borders, like Minneapolis into Wisconsin or New York City into effectively every adjacent state. Some overlap while others are isolated, especially in the west.

main

In the end, this is not a definitive way to look at geography within the Lower 48, but it does start to push the observer to rethink conventional regions of influence and defined borders. From the abstract: “The emergence in the United States of large-scale ‘megaregions’ centered on major metropolitan areas is a phenomenon often taken for granted in both scholarly studies and popular accounts of contemporary economic geography. We compare a method which uses a visual heuristic for understanding areal aggregation to a method which uses a computational partitioning algorithm, and we reflect upon the strengths and limitations of both. We discuss how choices about input parameters and scale of analysis can lead to different results, and stress the importance of comparing computational results with ‘common sense’ interpretations of geographic coherence.”

Share on Facebook





[ By WebUrbanist in Culture & History & Travel. ]

[ WebUrbanist | Archives | Galleries | Privacy | TOS ]


WebUrbanist

 
Comments Off on US Megaregions: Algorithm Redefines Boundaries of Metropolitan Areas

Posted in Creativity

 

Google algorithm can caption images with 93.9 % accuracy

27 Sep

Anyone who manages a large image library knows how important keywording and captioning are for categorizing and keeping things searchable. They also know how time-consuming these tasks can be. That’s where artificial intelligence may be able to lend a hand though, and the updated version of Google’s trainable ‘Show and Tell’ algorithm, which has just been made open source, is now capable of describing the contents of an image with an impressive 93.9% accuracy.

Google’s model generates a new captions by using concepts learned from pre-captioned images in the training set.

According to an article on the Google Research Blog the updated algorithm is faster to train and produces more detailed descriptions. The Google researchers trained ‘Show and Tell’ by showing it pre-captioned images of a specific scene to teach it to accurately caption similar scenes without any human help. By making ‘Show and Tell’ open source Google aims to promote research in the field of image recognition.

After the update the image model is now capable of providing more detailed descriptions and more likely to include color descriptions.

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Google algorithm can caption images with 93.9 % accuracy

Posted in Uncategorized

 

Corel AfterShot Pro 3 launches with new touchup tool, recovery algorithm

12 May

Corel has launched AfterShot Pro 3, the latest version of its photo editing software. The newest version brings several added features and updates, including a Lens Correction Development Kit for creating custom lens corrections, an in-app plugin manager, and a few new and improved tools for touching up photos.

AfterShot Pro 3 is equipped with a completely new Highlight Recovery algorithm, and as such Corel claims the Highlight Recovery Range slider can pull more tones and details from overexposed Raw photos. Joining the new algorithm is the addition of ‘comprehensive watermarking,’ including the ability to watermark in batches, rotate the watermark’s angle, adjust its size, and alter its transparency. 

Another new editing tool is Blemish Removal & Correction, which aims to eliminate the need to use a separate app like Photoshop to remove blemishes and perform other touchups and small corrections. Photo presets can also be applied via the new in-app preset library; both premium and free presets are offered.

Finally, AfterShot Pro 3 features a new modular delivery system for providing updated and new Raw profiles more quickly than the previous software version. With this, new camera profiles are available to download in-app as soon as they’re released by the company’s development team.

Corel AfterShot Pro 3 is available in English, German and Japanese through the product’s website; Windows, Mac and Linux are supported. The price for new customers is $ 79.99 USD/CAD, while existing customers can upgrade to the newest version for $ 59.99 USD/CAD.

Via: MarketWired

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Corel AfterShot Pro 3 launches with new touchup tool, recovery algorithm

Posted in Uncategorized

 

Instagram is changing its feed to use algorithm

18 Mar

Instagram has confirmed in a blog post that it will change the way the photos and videos of your friends are displayed in its stream. Chronological sorting will be scrapped in favor of a new algorithm that sorts images based on ‘the likelihood that you’ll be interested in the content, your relationship with the person posting and the timeliness of the post.’ Read more

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Instagram is changing its feed to use algorithm

Posted in Uncategorized

 

MIT algorithm aims to eradicate reflections from photos taken through windows

14 May

Researchers at the Massachusetts Institute of Technology claim to have developed a method for eliminating reflections in glass via digital processing. It is hoped that with further development the idea could see its way into digital cameras, allowing reflections to be automatically removed when they interfere with the view through a window. Read more

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on MIT algorithm aims to eradicate reflections from photos taken through windows

Posted in Uncategorized

 

Google develops ‘coherent’ image identification algorithm

20 Nov

Google is working on an image identification technology at its Research Labs in Mountain View, California. The latest complex algorithm from the search engine giant is able to systematically ‘produce captions to accurately describe images the first time it sees them’, creating coherent sentences rather than individual tags. Read more

Articles: Digital Photography Review (dpreview.com)

 
Comments Off on Google develops ‘coherent’ image identification algorithm

Posted in Uncategorized