主要内容

投影

使用主成分分析 (PCA) 对网络层进行投影;减少可学习参数的数量

对层进行投影,先使用代表训练数据的数据集对层激活执行主成分分析 (PCA),然后对层的可学习参数应用线性投影。当您使用无库的 C/C++ 代码生成将网络部署到嵌入式硬件时,投影的深度神经网络的前向传导通常会更快。

有关 Deep Learning Toolbox™ Model Compression Library 中提供的压缩技术的详细概述,请参阅Reduce Memory Footprint of Deep Neural Networks

Simplified illustration of projection. On the left is a sketch of a neural network with two layers that consist of three and two neurons, respectively. Every neuron in the first layer is connected to every neuron in the second layer. An arrow points to a second sketch on the right, which shows a different model that consists of three layers, with three, one, and two neurons, respectively. The right network has fewer weights in total compared to the left.

函数

compressNetworkUsingProjectionCompress neural network using projection (自 R2022b 起)
neuronPCAPrincipal component analysis of neuron activations (自 R2022b 起)
unpackProjectedLayersUnpack projected layers of neural network (自 R2023b 起)
ProjectedLayerCompressed neural network layer using projection (自 R2023b 起)
gruProjectedLayerGated recurrent unit (GRU) projected layer for recurrent neural network (RNN) (自 R2023b 起)
lstmProjectedLayerLong short-term memory (LSTM) projected layer for recurrent neural network (RNN) (自 R2022b 起)

主题

精选示例