• Author(s): Jinuk Kim, Marwa El Halabi, Mingi Ji, Hyun Oh Song

“LayerMerge: Neural Network Depth Compression through Layer Pruning and Merging” introduces a novel approach to reducing the depth of neural networks while maintaining their performance. This method, named Layer Merge, focuses on compressing neural networks by pruning and merging layers, which can significantly reduce the computational complexity and memory requirements of deep learning models.

Layer merge addresses the challenge of neural network compression by identifying and removing redundant layers and merging similar ones. The process begins with a thorough analysis of the network’s structure to detect layers that contribute minimally to the overall performance. These layers are then pruned, effectively reducing the depth of the network. Following this, the method identifies layers with similar functionalities and merges them, further compressing the network without compromising its accuracy.

The key innovation of Layer Merge lies in its ability to maintain the performance of the original network while achieving significant compression. This is accomplished through a combination of advanced pruning techniques and a novel merging algorithm that ensures the merged layers retain the essential features of the original layers. The result is a more compact and efficient neural network that requires less computational power and memory.

The paper provides extensive experimental results to demonstrate the effectiveness of layer merge. These results include quantitative evaluations on standard benchmarks, showing that the compressed networks achieve comparable or even superior performance to their original counterparts. Additionally, the authors present qualitative analyses that highlight the practical benefits of the method, such as reduced inference time and lower energy consumption. One of the key advantages of Layer Merge is its versatility. The method can be applied to a wide range of neural network architectures, making it a valuable tool for various applications, from image recognition to natural language processing. By reducing the depth of neural networks, Layer Merge enables the deployment of deep learning models on resource-constrained devices, such as mobile phones and embedded systems.

“Layer Merge: Neural Network Depth Compression through Layer Pruning and Merging” presents a significant advancement in the field of neural network compression. By effectively pruning and merging layers, the method achieves substantial reductions in network depth while maintaining performance. This research has important implications for the development of more efficient and scalable deep learning models, with potential applications in numerous domains requiring high-performance neural networks.