24/7 Space News
INTERNET SPACE
FeatUp: Revolutionizing Computer Vision with High-Resolution Feature Analysis
FeatUp is an algorithm that upgrades the resolution of deep networks for improved performance in computer vision tasks such as object recognition, scene parsing, and depth measurement. Credits:Image: Mark Hamilton and Alex Shipps/MIT CSAIL, top image via Unsplash.
ADVERTISEMENT
The 2024 Humans To Mars Summit - May 07-08, 2024 - Washington D.C.
FeatUp: Revolutionizing Computer Vision with High-Resolution Feature Analysis
by Rachel Gordon | MIT CSAIL
Boston MA (SPX) Mar 21, 2024

Imagine yourself glancing at a busy street for a few moments, then trying to sketch the scene you saw from memory. Most people could draw the rough positions of the major objects like cars, people, and crosswalks, but almost no one can draw every detail with pixel-perfect accuracy. The same is true for most modern computer vision algorithms: They are fantastic at capturing high-level details of a scene, but they lose fine-grained details as they process information.

Now, MIT researchers have created a system called "FeatUp" that lets algorithms capture all of the high- and low-level details of a scene at the same time - almost like Lasik eye surgery for computer vision.

When computers learn to "see" from looking at images and videos, they build up "ideas" of what's in a scene through something called "features." To create these features, deep networks and visual foundation models break down images into a grid of tiny squares and process these squares as a group to determine what's going on in a photo. Each tiny square is usually made up of anywhere from 16 to 32 pixels, so the resolution of these algorithms is dramatically smaller than the images they work with. In trying to summarize and understand photos, algorithms lose a ton of pixel clarity.

The FeatUp algorithm can stop this loss of information and boost the resolution of any deep network without compromising on speed or quality. This allows researchers to quickly and easily improve the resolution of any new or existing algorithm. For example, imagine trying to interpret the predictions of a lung cancer detection algorithm with the goal of localizing the tumor. Applying FeatUp before interpreting the algorithm using a method like class activation maps (CAM) can yield a dramatically more detailed (16-32x) view of where the tumor might be located according to the model.

FeatUp not only helps practitioners understand their models, but also can improve a panoply of different tasks like object detection, semantic segmentation (assigning labels to pixels in an image with object labels), and depth estimation. It achieves this by providing more accurate, high-resolution features, which are crucial for building vision applications ranging from autonomous driving to medical imaging.

"The essence of all computer vision lies in these deep, intelligent features that emerge from the depths of deep learning architectures. The big challenge of modern algorithms is that they reduce large images to very small grids of 'smart' features, gaining intelligent insights but losing the finer details," says Mark Hamilton, an MIT PhD student in electrical engineering and computer science, MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) affiliate, and a co-lead author on a paper about the project.

"FeatUp helps enable the best of both worlds: highly intelligent representations with the original image's resolution. These high-resolution features significantly boost performance across a spectrum of computer vision tasks, from enhancing object detection and improving depth prediction to providing a deeper understanding of your network's decision-making process through high-resolution analysis."

Resolution renaissance
As these large AI models become more and more prevalent, there's an increasing need to explain what they're doing, what they're looking at, and what they're thinking.

But how exactly can FeatUp discover these fine-grained details? Curiously, the secret lies in wiggling and jiggling images.

In particular, FeatUp applies minor adjustments (like moving the image a few pixels to the left or right) and watches how an algorithm responds to these slight movements of the image. This results in hundreds of deep-feature maps that are all slightly different, which can be combined into a single crisp, high-resolution, set of deep features.

"We imagine that some high-resolution features exist, and that when we wiggle them and blur them, they will match all of the original, lower-resolution features from the wiggled images. Our goal is to learn how to refine the low-resolution features into high-resolution features using this 'game' that lets us know how well we are doing," says Hamilton.

This methodology is analogous to how algorithms can create a 3D model from multiple 2D images by ensuring that the predicted 3D object matches all of the 2D photos used to create it. In FeatUp's case, they predict a high-resolution feature map that's consistent with all of the low-resolution feature maps formed by jittering the original image.

The team notes that standard tools available in PyTorch were insufficient for their needs, and introduced a new type of deep network layer in their quest for a speedy and efficient solution. Their custom layer, a special joint bilateral upsampling operation, was over 100 times more efficient than a naive implementation in PyTorch.

The team also showed this new layer could improve a wide variety of different algorithms including semantic segmentation and depth prediction. This layer improved the network's ability to process and understand high-resolution details, giving any algorithm that used it a substantial performance boost.

"Another application is something called small object retrieval, where our algorithm allows for precise localization of objects. For example, even in cluttered road scenes algorithms enriched with FeatUp can see tiny objects like traffic cones, reflectors, lights, and potholes where their low-resolution cousins fail. This demonstrates its capability to enhance coarse features into finely detailed signals," says Stephanie Fu '22, MNG '23, a PhD student at the University of California at Berkeley and another co-lead author on the new FeatUp paper.

"This is especially critical for time-sensitive tasks, like pinpointing a traffic sign on a cluttered expressway in a driverless car. This can not only improve the accuracy of such tasks by turning broad guesses into exact localizations, but might also make these systems more reliable, interpretable, and trustworthy."

What next?
Regarding future aspirations, the team emphasizes FeatUp's potential widespread adoption within the research community and beyond, akin to data augmentation practices. "The goal is to make this method a fundamental tool in deep learning, enriching models to perceive the world in greater detail without the computational inefficiency of traditional high-resolution processing," says Fu.

"FeatUp represents a wonderful advance towards making visual representations really useful, by producing them at full image resolutions," says Cornell University computer science professor Noah Snavely, who was not involved in the research.

"Learned visual representations have become really good in the last few years, but they are almost always produced at very low resolution - you might put in a nice full-resolution photo, and get back a tiny, postage stamp-sized grid of features. That's a problem if you want to use those features in applications that produce full-resolution outputs. FeatUp solves this problem in a creative way by combining classic ideas in super-resolution with modern learning approaches, leading to beautiful, high-resolution feature maps."

"We hope this simple idea can have broad application. It provides high-resolution versions of image analytics that we'd thought before could only be low-resolution," says senior author William T. Freeman, an MIT professor of electrical engineering and computer science professor and CSAIL member.

Lead authors Fu and Hamilton are accompanied by MIT PhD students Laura Brandt SM '21 and Axel Feldmann SM '21, as well as Zhoutong Zhang SM '21, PhD '22, all current or former affiliates of MIT CSAIL. Their research is supported, in part, by a National Science Foundation Graduate Research Fellowship, by the National Science Foundation and Office of the Director of National Intelligence, by the U.S. Air Force Research Laboratory, and by the U.S. Air Force Artificial Intelligence Accelerator. The group will present their work in May at the International Conference on Learning Representations.

Research Report:FeatUp: A Model-Agnostic Framework for Features at any Resolution

Related Links
Computer Science and Artificial Intelligence Laboratory (CSAIL)
Satellite-based Internet technologies

Subscribe Free To Our Daily Newsletters

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
INTERNET SPACE
Apple in talks to bring Google's Gemini AI to iPhone: report
San Francisco (AFP) Mar 18, 2024
Apple is in advanced negotiations with Google to adopt its powerful Gemini AI models for the iPhone, Bloomberg reported on Monday. The move, if confirmed and resulting in a deal, would answer one of the big questions about the iPhone-maker's AI strategy, which has been criticized as lagging its tech giant rivals. Google's share price soared by well over six percent on the news. Apple's was up by as much as 2.5 percent. According to Bloomberg, Apple is seeking to partner with Google for its n ... read more

ADVERTISEMENT
ADVERTISEMENT
INTERNET SPACE
Imagining sustainability through the eyes of future generations

Xi tells Dutch PM Rutte 'no force can stop' China tech progress

Spacecraft with first Belarussian woman cosmonaut takes off

Artemis II's Orion Spacecraft to Undergo Critical Manual Handling Test

INTERNET SPACE
RocketStar unveils fusion-enhanced electric thruster for spacecraft

Final Voyage Delayed: ULA Postpones Last Delta IV Heavy Launch with Classified NRO Payload

Artemis IV Mission Advances with Completion of SLS Payload Adapter Testing

ISRO's RLV-LEX-02 Mission: A Leap Forward in Autonomous Space Vehicle Landing

INTERNET SPACE
Curiosity's Encore Journey Along Upper Gediz Vallis Ridge

A Return to Your Regularly Scheduled Touch-And-Go: Sols 4130-4131

NASA and JAXA advance Martian Moons study with instrument handoff

It's go for drilling at Mineral King: Sols 4125-4126:

INTERNET SPACE
Shenzhou 17 astronauts complete China's first in-space repair job

Tiangong Space Station's Solar Wings Restored After Spacewalk Repair by Shenzhou XVII Team

BIT advances microbiological research on Chinese Space Station

Chang'e 6 and new rockets highlight China's packed 2024 space agenda

INTERNET SPACE
Airbus Prepares EUTELSAT 36D Satellite for Launch with Innovative BelugaST Transport

A New Dawn in Satellite Technology: MDA Space Unveils AURORA

Aerospace Corp., Space Foundation partner to boost inclusivity in space industry

LizzieSat signals success for Sidus Space in post-launch phase

INTERNET SPACE
UC San Diego Scientists Unveil Plant-Based Polymers that Biodegrade Microplastics in Months

Frost-resistant concrete technology from Drexel could make salt and shovels obsolete

Using nature's recipe for 3D-printed wood

New Insights into the Van Allen Belts: Rethinking Particle Dynamics in Space

INTERNET SPACE
Unveiling hydrogen's role in life's early energy mechanisms

Life Detection on Ice Moons Could Be Within Reach, New Study Shows

Loathed by scientists, loved by nature: sulfur and the origin of life

Webb finds ethanol, other icy ingredients for making planets

INTERNET SPACE
Unlocking the Secrets of Eternal Ice in the Kuiper Belt

Hubble's Latest Gaze Reveals Jupiter's Dynamic Weather Patterns

NASA Armstrong Updates 1960s Concept to Study Giant Planets

NASA's Europa Jupiter Mission will be packed with humanity's messages

Subscribe Free To Our Daily Newsletters


ADVERTISEMENT



The content herein, unless otherwise known to be public domain, are Copyright 1995-2023 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.