Neural networks - and deep neural networks in particular - are driving machine learning that is closer than ever to the thought processes of the human brain. Their ability to learn and drive decisions from data makes them ideal candidates to use the information gathered by edge devices.
These edge devices, from smartphones to IoT sensors, can collect so much data for neural networks that a level of refinement and focus is. This process is called deep neural network pruning, and this blog explores how it works and why it’s so beneficial.
Deep neural network pruning is the act of removing parameters from neural network models that aren’t important to its successful operation. The theory mimics the principle that the brain turns down use of connections between neurons that aren’t needed, to devote more energy to the more important ones. In practice, this means reducing the weights (and not the biases) of the parameters, in order to reduce the size of the Generative AI model for example, and make its function and findings more efficient.
In the context of edge devices, deep neural network pruning is important because those devices tend to be relatively constrained in terms of their capacity across processing, latency, memory and network bandwidth for model deployment. Filtering unnecessary workloads ensures the most important data processing takes place, maximizing the efficiency and results of the model.
Edge devices process data closer to where it’s generated, rather than sending additional workload to central servers or cloud-based data centers. These devices include (and are by no means limited to) smartphones, IoT devices like cameras and thermostats, and even microcontrollers in modern cars.
These devices are extremely beneficial for modern computing architectures. They can reduce demands on bandwidth and latency, support greater insights and real-time decision-making, cut energy consumption, and improve data privacy and security. But as mentioned in the previous section, edge devices often aren’t suited to the demands of neural networks like high storage space volumes, intensive compute power and large memory requirements.
With the help of deep neural network pruning, therefore, edge devices can support important functions based around neural networks, including face and speech recognition, autonomous vehicles, healthcare wearables, augmented reality and more.
Deep neural network pruning enables a significant reduction in the size of neural networks, which in turn reduces the demand for memory and storage. This approach can reduce model weights by more than 50% (and in our experience, can deliver reductions of up to 80%) while maintaining a less than 1% drop in accuracy. Not only does this relieve the pressure on edge devices, but it also reduces the cost and demand on requirements at the center of the network, whether in the cloud or on-premise.
In turn, this makes neural networks more practical to deploy in a wider range of settings where RAM and flash storage are limited. It also speeds up execution time through faster inference and lower latency, which is especially critical in applications like autonomous vehicles or AR which operate in real-time.
The use of machine learning algorithms, and artificial intelligence more widely, is coming under increasing social and regulatory scrutiny for its impact on the environment. Processing and transferring large volumes of data come with energy and emissions implications in training and inference, which deep neural network pruning can address by stripping out unnecessary workloads.
Ensuring AI runs on edge devices rather than on cloud platforms supports this further by cutting out more data transmissions and lowering the energy consumption of the cloud, simultaneously lowering the cost of AI in the process.
If edge devices are able to complete AI tasks independently, then the need for a fast and stable Internet connection to transfer data is removed. This is especially useful in supporting functionality offline: for example, face and image recognition and real-time processing can take place in remote and rural areas. The latency and lead time of waiting for responses from the cloud can also be removed.
Additionally, privacy and security can be enhanced as sensitive data (such as biometric information and medical records) can stay on the device in question, minimizing its potential attack surface. This can be instrumental in not only securing that data, but also in maintaining compliance with key privacy regulations like GDPR and CCPA.
Deep neural network pruning is already being applied to good effect in a range of industries where neural networks are supported by edge devices.
Getting deep neural network pruning right is important to maximize the benefit of edge devices and the insights they can generate for neural networks. It may sound very complicated, but there are several ways to do it adequately, including:
If you feel you need support in weight pruning, then the help and advice of an expert AI and machine learning partner can be invaluable. Explore the Ciklum approach to data and AI today, then contact our team to find out more.