Computer

The Impact of Network Pruning on Deep Learning Models

The Impact of Network Pruning on Deep Learning Models

Model pruning is the practice of removing weights that do not improve the performance of a model. We can compress and deploy our workhorse neural networks onto mobile phones and other resource-constrained devices thanks to careful pruning.

Researchers in computer science have shown that a widely used technique known as neural network pruning can degrade the performance of deep learning models, detailed what causes these performance issues, and demonstrated a solution to the problem.

Deep learning is a type of artificial intelligence that can classify objects such as images, text, or sound. It can, for example, be used to identify people based on facial images. Deep learning models, on the other hand, frequently necessitate a large amount of computing power to run. When a deep learning model is used for some applications, this presents challenges.

To address these challenges, some systems engage in “neural network pruning.” This effectively makes the deep learning model more compact and, therefore, able to operate while using fewer computing resources.

I think the most important aspect of this work is that we now have a more thorough understanding of exactly how network pruning can influence the performance of deep learning models to identify minority groups, both theoretically and empirically.

Jung-Eun Kim

“However, our research shows that this network pruning can impair the ability of deep learning models to identify some groups,” says Jung-Eun Kim, co-author of a paper on the work and an assistant professor of computer science at North Carolina State University.

“For example, if a security system uses deep learning to scan people’s faces in order to determine whether they have access to a building, the deep learning model would have to be made compact so that it can operate efficiently. This may work fine most of the time, but the network pruning could also affect the deep learning model’s ability to identify some faces.”

In their new paper, the researchers lay out why network pruning can adversely affect the performance of the model at identifying certain groups which the literature calls “minority groups” and demonstrate a new technique for addressing these challenges.

The-Impact-of-Network-Pruning-on-Deep-Learning-Models-1
How network pruning can skew deep learning models

Two factors explain how network pruning can impair the performance of deep learning models.

In technical terms, these two factors are: disparity in gradient norms across groups; and disparity in Hessian norms associated with inaccuracies of a group’s data. In practical terms, this means that deep learning models can become less accurate in recognizing specific categories of images, sounds or text. Specifically, the network pruning can amplify accuracy deficiencies that already existed in the model.

For example, if a deep learning model is trained to recognize faces using a data set that includes the faces of 100 white people and 60 Asian people, it might be more accurate at recognizing white faces, but could still achieve adequate performance for recognizing Asian faces. After network pruning, the model is more likely to be unable to recognize some Asian faces.

“The deficiency may not have been noticeable in the original model, but because it’s amplified by the network pruning, the deficiency may become noticeable,” Kim says.

“To mitigate this problem, we’ve demonstrated an approach that uses mathematical techniques to equalize the groups that the deep learning model is using to categorize data samples,” Kim says. “In other words, we are using algorithms to address the gap in accuracy across groups.”

In testing, the researchers demonstrated that using their mitigation technique improved the fairness of a deep learning model that had undergone network pruning, essentially returning it to pre-pruning levels of accuracy.

“I think the most important aspect of this work is that we now have a more thorough understanding of exactly how network pruning can influence the performance of deep learning models to identify minority groups, both theoretically and empirically,” Kim says. “We’re also open to working with partners to identify unknown or overlooked impacts of model reduction techniques, particularly in real-world applications for deep learning models.”