Movatterモバイル変換


[0]ホーム

URL:


CN112418424A - A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio - Google Patents

A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio
Download PDF

Info

Publication number
CN112418424A
CN112418424ACN202011462942.3ACN202011462942ACN112418424ACN 112418424 ACN112418424 ACN 112418424ACN 202011462942 ACN202011462942 ACN 202011462942ACN 112418424 ACN112418424 ACN 112418424A
Authority
CN
China
Prior art keywords
zero
blocks
sparse
sparse coding
pruning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011462942.3A
Other languages
Chinese (zh)
Inventor
李文斌
何爱清
刘潇
霍静
姚丽丽
高阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Nanjing University
Original Assignee
Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd, Nanjing UniversityfiledCriticalJiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Priority to CN202011462942.3ApriorityCriticalpatent/CN112418424A/en
Publication of CN112418424ApublicationCriticalpatent/CN112418424A/en
Pendinglegal-statusCriticalCurrent

Links

Images

Classifications

Landscapes

Abstract

Translated fromChinese

本发明提出了一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,包括:应用剪枝技术对超参数化的DNN模型进行稀疏化处理;提出一种分层稀疏编码方法LSC,通过设计元数据的编码机制来提高修剪后的DNN模型的压缩率;设计一种多过程解码机制使得无需完全解码即支持矩阵运算,节省运行内存;所述分层稀疏编码方法通过减少元数据量来最大化压缩率,包括:将压缩过程分为块层和编码层;在块层中,将稀疏矩阵划分为多个小块,然后删除零值块;在编码层中,提出了一种新颖的带标记相对索引方法SRI,来进一步编码这些非零值块;所述多过程解码机制在推断阶段加快编码矩阵的乘法。最后通过实验对比证明所提出的LSC方法相对于其他稀疏编码方法的有效性。

Figure 202011462942

The invention proposes a layered sparse coding method for a pruned deep neural network with extremely high compression ratio, including: applying pruning technology to sparse a hyperparameterized DNN model; and providing a layered sparse coding method LSC, improves the compression rate of the pruned DNN model by designing a metadata encoding mechanism; designs a multi-process decoding mechanism to support matrix operations without full decoding, saving running memory; the layered sparse coding method reduces the metadata The amount of data to maximize the compression rate includes: dividing the compression process into a block layer and an encoding layer; in the block layer, the sparse matrix is divided into multiple small blocks, and then the zero-valued blocks are deleted; A novel labeled relative indexing method SRI is used to further encode these non-zero valued blocks; the multi-pass decoding mechanism speeds up the multiplication of the encoding matrix at the inference stage. Finally, the effectiveness of the proposed LSC method compared with other sparse coding methods is proved by experimental comparison.

Figure 202011462942

Description

Layered sparse coding method of pruning deep neural network with extremely high compression ratio
Technical Field
The invention relates to a hierarchical sparse coding method of a pruning deep neural network with extremely high compression ratio, belonging to the field of machine learning.
Background
Deep neural networks DNN have evolved as state-of-the-art in many areas, particularly in the areas of computer vision, natural language processing, and audio processing. However, the large growth of hidden layers and neurons consumes considerable hardware storage and memory bandwidth, which presents a serious challenge for many resource-constrained scenarios in real-world applications. Especially the advent of the post-morgan era has slowed hardware replacement cycles. Specifically, there are two major bottlenecks to the current DNN: 1. in conflict with resource-constrained application platforms, such as automatic driving tools, mobile phones, mobile robots and augmented reality AR, which are very sensitive to the energy consumption and the amount of calculation of the DNN model, there is an urgent need for a DNN model with low energy consumption but good performance. 2. In conflict with new accelerators such as FPGAs, custom ASICs and AI specific chips, which are powerful computational accelerators for DNN but are also sensitive to hardware storage, memory bandwidth and parallelism, it is clear that DNN is expected to reduce hardware storage and memory usage while enjoying parallelism.
In order to solve the above-mentioned bottleneck, many compression methods, such as pruning and quantization, are proposed, and the number of weights of the trained DNN model can be easily reduced by using these effective compression methods. For example, the weight matrix of the target network can be made very sparse by using a classical magnitude-based pruning method, in order to store and migrate these sparse weights, they are usually decomposed into two types of data, namely non-zero weights and metadata, and these non-zero weights are encoded and decoded by using metadata expression index information, so that when an acceptable final performance is ensured, a high compression ratio can be achieved by reducing the number of non-zero weights as much as possible; however, encoding non-zero weights requires a large amount of metadata, which is several times more than the number of actual non-zero weights.
In summary, a large amount of metadata is that the weights of the pruning DNN compression, storage and migration obstacle learning compression DNN model become sparse after pruning, so how to store and migrate these sparse weights has become a focus of attention in recent years. Generally, these studies can be divided into two categories, depending on the objective: compression ratio and parallel computation.
Most compression methods use some classical sparse coding methods such as MATLAB, TensorFlow and Pytorch to integrate COO methods into default sparse coding methods, while Scipy encodes sparse matrices using compressed sparse rows/columns, CSR/CSC for short, based only on the programming framework they use. New methods such as bitmasks and relative indices have recently been proposed which are capable of encoding sparse models, but they are procedural methods which are difficult to implement in parallel. In order to fully utilize the resources of the deep learning accelerator, a series of novel sparse coding methods including block compression sparse column BCSC and nested bit mask NB have been proposed in recent years, and these methods are suitable for parallelization environment, but the compressed model still consumes a large amount of storage and memory bandwidth.
A significant challenge with both of the above methods is that it is difficult to achieve both high compression ratio and efficient calculation. Unlike these sparse coding methods, the present invention not only allows parallel model inference, but also requires very little metadata, thus providing higher compression rates for pruned deep models.
Disclosure of Invention
The purpose of the invention is as follows: a hierarchical sparse coding method of a pruning deep neural network with a very high compression ratio is provided to solve the problems in the prior art.
The technical scheme is as follows: a hierarchical sparse coding method of a pruning deep neural network with a very high compression ratio specifically comprises the following steps:
step 1, initializing and training an over-parameterized DNN model;
step 2, pruning the model to make the model as sparse as possible;
step 3, adopting a layered sparse coding LSC method to further compress and code the sparse weights, applying a Block bit mask Mechanism Block bitmap Mechanism to divide a sparse matrix into a plurality of small blocks for each sparse weight matrix in a Block layer, stretching and splicing all nonzero-value blocks into a vector, and sending the flattened vector into a subsequent coding layer;
step 4, encoding the flattened vector by using the extreme finite element data through an SRI method with a mark relative index at an encoding layer, and executing compression encoding;
and step 5, an inference stage, namely starting a matrix multiplication process as early as possible by using an intermediate decoding result to realize high-performance calculation.
The general procedure of learning the compressed DNN model ofstep 1 andstep 2 is as follows:
step 11, initializing and training an over-parameterized DNN model;
step 12, eliminating the weight which contributes less to prediction through pruning, and retraining the model;
and step 13, repeating the pruning and training processes for a plurality of times, wherein the finally obtained model keeps the performance similar to the original model but has less effective weight.
The hierarchical sparse coding LSC method of thestep 3 further comprises the following steps:
step 31, a block bit mask mechanism is adopted to divide each sparse weight matrix in the block layer into a plurality of small blocks, and for each block, if any non-zero weight exists, a 1-bit signal is used to mark the block as true, otherwise, the block is marked as false, so that a mask consisting of a plurality of 1-bit marks is obtained, and all blocks can be marked;
step 32, flattening all the non-zero blocks into a vector, and inputting the flattened vector into the coding layer.
The LSC method in step S4 adopts SRI method to perform high-intensity compression on the flattened non-zero block by its coding layer, and compresses and codes all the non-zero weights.
The inference process of step S5 is further:
step 51, calculating tree structure: the calculation tree is used for determining the calculation process of matrix multiplication, and firstly, the multiplication of two matrixes W multiplied by X is decomposed into the calculation of a plurality of sub-matrixes; according to the principle of block matrix multiplication, dividing W into a plurality of sub-matrices, then converting W multiplied by X into calculation among the sub-matrices, and further converting the calculation process of each row into a calculation tree;
step 52, pruning the calculation tree: the high sparsity of the DNN after pruning leads to a large number of zero-value blocks, multiplication of the zero blocks can be skipped, and a calculation tree is pruned according to the marks of the zero-value blocks by a block bit mask mechanism;
step 53, SRI decoding and submatrix multiplication: the SRI decoding process restores the SRI code into non-zero blocks, and the sub-matrix multiplication process adopts the non-zero blocks to execute sub-matrix multiplication; the two processes are relatively independent, once the multiplication of the submatrix is completed, the decoded non-zero block is destroyed, and the storage bandwidth is saved;
step 54, intermediate result integration: all intermediate calculation results of the sub-matrix multiplication are accumulated to obtain a final result and can be implemented in parallel.
Has the advantages that: the method has a novel and simple hierarchical sparse coding structure, provides a novel SRI method, can code the non-zero weight by using the minimum space, and designs an effective decoding mechanism for the provided LSC method so as to accelerate the multiplication of a coding matrix in an inference stage; a large number of comparison experiment results are analyzed, so that the LSC method provided by the invention obtains considerable benefits in DNN compression and inference calculation of pruning, and the yield exceeds the leading level of the field.
Drawings
Fig. 1 is a structural diagram of a hierarchical sparse coding LSC method.
Fig. 2 is a schematic diagram comparing a relative indexing method with marks and a conventional relative indexing method.
FIG. 3 is a diagram of a computation tree structure.
Fig. 4 is a matrix multiplication diagram under the LSC method.
Detailed Description
In an alternative embodiment, as shown in fig. 1, a hierarchical sparse coding method for pruning deep neural networks with extremely high compression ratio includes maximizing compression ratio by greatly reducing metadata amount, optimizing coding and inference process design of sparse matrix using a novel and effective hierarchical sparse coding framework; a sparse coding method with extremely high compression ratio is provided, wherein an LSC method is used, the LSC method comprises two key layer block layers and a coding layer, in the block layers, a sparse matrix is divided into a plurality of small blocks, then zero blocks are deleted, and in the coding layer, a novel SRI method is provided for further coding the non-zero blocks; in addition, the invention designs an effective decoding mechanism for the LSC method to accelerate the multiplication of the coding matrix in the inference stage.
The method specifically comprises the following steps:
step 1, initializing and training an over-parameterized DNN model;
step 2, pruning the model to make the model as sparse as possible;
step 3, adopting a layered sparse coding LSC method to further compress and code the sparse weights, applying a Block bit mask Mechanism Block bitmap Mechanism to divide a sparse matrix into a plurality of small blocks for each sparse weight matrix in a Block layer, stretching and splicing all nonzero-value blocks into a vector, and sending the flattened vector into a subsequent coding layer;
step 4, encoding the flattened vector by using the extreme finite element data through an SRI method with a mark relative index at an encoding layer, and executing compression encoding;
and step 5, an inference stage, namely starting a matrix multiplication process as early as possible by using an intermediate decoding result to realize high-performance calculation.
In a further embodiment, the general procedure of learning the compressed DNN model ofstep 1 andstep 2 is as follows:
step 11, initializing and training an over-parameterized DNN model;
step 12, eliminating the weight which contributes less than an expected value to the prediction through pruning, and retraining the model;
and step 13, repeating the pruning and training processes for a plurality of times, wherein the finally obtained model keeps the performance similar to the original model but has less effective weight.
In a further embodiment, the hierarchical sparse coding LSC method ofstep 3 further includes:
step 31, a block bit mask mechanism is adopted to divide each sparse weight matrix in the block layer into a plurality of small blocks, and for each block, if any non-zero weight exists, a 1-bit signal is used to mark the block as true, otherwise, the block is marked as false, so that a mask consisting of a plurality of 1-bit marks is obtained, and all blocks can be marked;
step 32, flattening all the non-zero blocks into a vector, and inputting the flattened vector into the coding layer.
In a further embodiment, the LSC method of step S4, which uses SRI method to perform high-intensity compression on the flattened non-zero block with its coding layer, compresses and encodes all the non-zero weights.
In a further embodiment, the inference process of step S5 is further:
step 51, calculating tree structure: the calculation tree is used for determining the calculation process of matrix multiplication, and firstly, the multiplication of two matrixes W multiplied by X is decomposed into the calculation of a plurality of sub-matrixes; according to the principle of block matrix multiplication, dividing W into a plurality of sub-matrices, then converting W multiplied by X into calculation among the sub-matrices, and further converting the calculation process of each row into a calculation tree;
step 52, pruning the calculation tree: the high sparsity of the DNN after pruning leads to a large number of zero-value blocks, multiplication of the zero blocks can be skipped, and a calculation tree is pruned according to the marks of the zero-value blocks by a block bit mask mechanism;
step 53, SRI decoding and submatrix multiplication: the SRI decoding process restores the SRI code into non-zero blocks, and the sub-matrix multiplication process adopts the non-zero blocks to execute sub-matrix multiplication; the two processes are relatively independent, once the multiplication of the submatrix is completed, the decoded non-zero block is destroyed, and the storage bandwidth is saved;
step 54, intermediate result integration: all intermediate calculation results of the sub-matrix multiplication are accumulated to obtain a final result and can be implemented in parallel.
Although the preferred embodiments of the present invention have been described in detail, the present invention is not limited to the details of the embodiments, and various equivalent modifications can be made within the technical spirit of the present invention, and the scope of the present invention is also within the scope of the present invention.

Claims (9)

Translated fromChinese
1.一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,包括如下步骤:1. a layered sparse coding method with a pruning deep neural network having an extremely high compression ratio, is characterized in that, comprises the steps:步骤S1:初始化并训练一个过度参数化的DNN模型;Step S1: Initialize and train an over-parameterized DNN model;步骤S2:对模型进行剪枝,使其尽可能稀疏;Step S2: prune the model to make it as sparse as possible;步骤S3:采用分层稀疏编码LSC方法来进一步压缩和编码这些稀疏权重,在块层中,对于每个稀疏权值矩阵,应用块位掩码机制,将稀疏矩阵划分为多个小块,然后将所有非零值块拉伸拼接为一个向量,扁平化矢量被送入后续编码层;Step S3: The layered sparse coding LSC method is used to further compress and encode these sparse weights. In the block layer, for each sparse weight matrix, a block bitmask mechanism is applied to divide the sparse matrix into multiple small blocks, and then All non-zero value blocks are stretched and spliced into a vector, and the flattened vector is sent to the subsequent encoding layer;步骤S4:在编码层,通过带标记相对索引方法SRI,用极有限元数据编码该扁平化矢量,执行压缩编码;Step S4: at the coding layer, by using the marked relative index method SRI, the flattened vector is coded with extremely finite element data, and compression coding is performed;步骤S5:推断阶段,利用中间解码结果来尽早启动矩阵乘法过程,实现高性能计算。Step S5: In the inference stage, the intermediate decoding result is used to start the matrix multiplication process as early as possible to realize high-performance computing.2.根据权利要求1所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述步骤S1和步骤S2的学习压缩DNN模型的过程进一步为:2. a kind of layered sparse coding method with the pruning deep neural network of extremely high compression ratio according to claim 1, is characterized in that, the process of the learning compression DNN model of described step S1 and step S2 is further:步骤S11、初始化并训练一个超参数化的DNN模型;Step S11, initialize and train a hyperparameterized DNN model;步骤S12、通过剪枝来消除对预测贡献较小的权重,并重新训练该模型;Step S12, eliminate the weights that contribute less to the prediction by pruning, and retrain the model;步骤S13、重复剪枝和训练过程数次,最终获得的模型保持与原始模型相似但有效权重却更少的性能。Step S13, repeating the pruning and training process several times, the finally obtained model maintains performance similar to the original model but with fewer effective weights.3.根据权利要求1所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述步骤S3的分层稀疏编码LSC方法进一步为:3. a kind of layered sparse coding method with the pruning deep neural network of extremely high compression ratio according to claim 1, is characterized in that, the layered sparse coding LSC method of described step S3 is further:步骤S31、采用块位掩码机制,将块层中每个稀疏权值矩阵划分为多个小块,对于每个块,如果有任何非零权重,则使用1位信号将其标记为true,否则将其标记为false,由此获得由许多1位标记组成的掩码,可以标记所有块;Step S31, using the block bit mask mechanism to divide each sparse weight matrix in the block layer into multiple small blocks, for each block, if there is any non-zero weight, use a 1-bit signal to mark it as true, Otherwise mark it as false, thus obtaining a mask consisting of many 1-bit marks that can mark all blocks;步骤S32、将所有这些非零块展平为一个矢量,并将展平的矢量输入编码层。Step S32: Flatten all these non-zero blocks into a vector, and input the flattened vector into the coding layer.4.根据权利要求1所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述步骤S4的LSC方法,采用SRI方法将其编码层对扁平化的非零块执行高强度压缩,压缩和编码所有非零权重。4. a kind of layered sparse coding method with the pruning deep neural network of extremely high compression ratio according to claim 1, is characterized in that, the LSC method of described step S4 adopts SRI method to flatten its coding layer to The transformed non-zero blocks perform high-intensity compression, compressing and encoding all non-zero weights.5.根据权利要求1所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述步骤S5的推断过程进一步为:5. a kind of layered sparse coding method with a pruning deep neural network with extremely high compression ratio according to claim 1, is characterized in that, the inference process of described step S5 is further:步骤51、计算树构造;Step 51, computing tree structure;步骤S52、修剪计算树;Step S52, pruning the computation tree;步骤S53、SRI解码和子矩阵乘法;Step S53, SRI decoding and sub-matrix multiplication;步骤S54、中间结果集成。Step S54, the intermediate results are integrated.6.根据权利要求5所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述计算树构造进一步为:计算树用于确定矩阵乘法的计算流程,首先将两个矩阵W×X的乘法分解为多个子矩阵的计算;根据块矩阵乘法的原理,将W拆分为多个子矩阵,然后将W×X转换为几个子矩阵之间的计算,每行的计算过程可以进一步转换为计算树。6. a kind of layered sparse coding method with a pruned deep neural network with extremely high compression ratio according to claim 5, is characterized in that, described computation tree structure is further: Computation tree is used to determine the computation of matrix multiplication In the process, the multiplication of two matrices W×X is first decomposed into the calculation of multiple sub-matrices; according to the principle of block matrix multiplication, W is split into multiple sub-matrices, and then W×X is converted into the calculation between several sub-matrices , the computation process of each row can be further transformed into a computation tree.7.根据权利要求5的推断过程,其特征在于,所述修剪计算树进一步为:修剪后的DNN的高度稀疏性导致大量的零值块,可以跳过这些零块的乘法,根据块位掩码机制对零值块的标记修剪计算树。7. The inference process according to claim 5, wherein the pruning calculation tree is further: the high sparsity of the pruned DNN results in a large number of zero-valued blocks, and the multiplication of these zero-valued blocks can be skipped, according to the block bit mask. The code mechanism prunes the computation tree for the markup of zero-valued blocks.8.根据权利要求5所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述SRI解码和子矩阵乘法进一步为:SRI解码过程将SRI码恢复为非零块,子矩阵乘法过程采用这些非零块来执行子矩阵乘法;这两个过程是相对独立的,一旦子矩阵乘法完成,解码的非零块将被破坏,节省存储带宽。8. a kind of layered sparse coding method with a pruned deep neural network with extremely high compression ratio according to claim 5, is characterized in that, described SRI decoding and sub-matrix multiplication are further: SRI decoding process restores SRI code For non-zero blocks, the sub-matrix multiplication process uses these non-zero blocks to perform sub-matrix multiplication; these two processes are relatively independent, once the sub-matrix multiplication is completed, the decoded non-zero blocks will be destroyed, saving memory bandwidth.9.根据权利要求5所述的一种具有极高压缩比的剪枝深度神经网络的分层稀疏编码方法,其特征在于,所述中间结果集成进一步为:对子矩阵乘法的所有中间计算结果进行累积以获得最终结果且可以并行实现。9. The layered sparse coding method of a pruned deep neural network with a very high compression ratio according to claim 5, wherein the intermediate result integration is further: all intermediate calculation results of sub-matrix multiplication Accumulation is done to get the final result and can be implemented in parallel.
CN202011462942.3A2020-12-112020-12-11 A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression RatioPendingCN112418424A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202011462942.3ACN112418424A (en)2020-12-112020-12-11 A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202011462942.3ACN112418424A (en)2020-12-112020-12-11 A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio

Publications (1)

Publication NumberPublication Date
CN112418424Atrue CN112418424A (en)2021-02-26

Family

ID=74776613

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202011462942.3APendingCN112418424A (en)2020-12-112020-12-11 A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio

Country Status (1)

CountryLink
CN (1)CN112418424A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113037669A (en)*2021-03-112021-06-25哈尔滨海能达科技有限公司Method and device for coding small blocks in satellite internet
CN113947185A (en)*2021-09-302022-01-18北京达佳互联信息技术有限公司Task processing network generation method, task processing device, electronic equipment and storage medium
CN114723033A (en)*2022-06-102022-07-08成都登临科技有限公司Data processing method, data processing device, AI chip, electronic device and storage medium
WO2023164950A1 (en)*2022-03-042023-09-07Intel CorporationMethod and apparatus for accelerating deep leaning inference based on hw-aware sparsity pattern
CN116796813A (en)*2022-03-102023-09-22聚晶半导体股份有限公司Deep neural network processing device, decompression method and compression method
WO2025143833A1 (en)*2023-12-272025-07-03세종대학교산학협력단Method and device for generating matrix index information, and method and device for processing matrix using matrix index information

Citations (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN101212510A (en)*2006-12-302008-07-02上海贝尔阿尔卡特股份有限公司Method and device for searching for user relevance information in multi-LAN environment in fixed network
CN107528077A (en)*2016-06-222017-12-29香港理工大学 Generating Energy from Pollutants
CN109245773A (en)*2018-10-302019-01-18南京大学A kind of decoding method based on block circulation sparse matrix neural network
US20190184561A1 (en)*2017-12-152019-06-20The Regents Of The University Of CaliforniaMachine Learning based Fixed-Time Optimal Path Generation
CN110191339A (en)*2019-05-222019-08-30上海富瀚微电子股份有限公司Rate estimation core cell, rate estimation device and code rate estimation method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN101212510A (en)*2006-12-302008-07-02上海贝尔阿尔卡特股份有限公司Method and device for searching for user relevance information in multi-LAN environment in fixed network
CN107528077A (en)*2016-06-222017-12-29香港理工大学 Generating Energy from Pollutants
US20190184561A1 (en)*2017-12-152019-06-20The Regents Of The University Of CaliforniaMachine Learning based Fixed-Time Optimal Path Generation
CN109245773A (en)*2018-10-302019-01-18南京大学A kind of decoding method based on block circulation sparse matrix neural network
CN110191339A (en)*2019-05-222019-08-30上海富瀚微电子股份有限公司Rate estimation core cell, rate estimation device and code rate estimation method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
LIU X等: "Layerwise Sparse Coding for Pruned Deep Neural Networks with Extreme Compression Ratio", 《PROCEEDINGS OF THE AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE》, vol. 34, no. 4, pages 4901 - 4903*
姜飞: "基于循环卷积的张量稀疏模型及应用研究", 《中国博士学位论文全文数据库 (信息科技辑)》, no. 1, pages 138 - 72*
常卫东: "智能入侵检测中的特征提取和集成学习技术研究", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》, no. 11, pages 139 - 171*

Cited By (8)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113037669A (en)*2021-03-112021-06-25哈尔滨海能达科技有限公司Method and device for coding small blocks in satellite internet
CN113037669B (en)*2021-03-112022-08-05哈尔滨海能达科技有限公司Method and device for coding small blocks in satellite internet
CN113947185A (en)*2021-09-302022-01-18北京达佳互联信息技术有限公司Task processing network generation method, task processing device, electronic equipment and storage medium
CN113947185B (en)*2021-09-302022-11-18北京达佳互联信息技术有限公司Task processing network generation method, task processing device, electronic equipment and storage medium
WO2023164950A1 (en)*2022-03-042023-09-07Intel CorporationMethod and apparatus for accelerating deep leaning inference based on hw-aware sparsity pattern
CN116796813A (en)*2022-03-102023-09-22聚晶半导体股份有限公司Deep neural network processing device, decompression method and compression method
CN114723033A (en)*2022-06-102022-07-08成都登临科技有限公司Data processing method, data processing device, AI chip, electronic device and storage medium
WO2025143833A1 (en)*2023-12-272025-07-03세종대학교산학협력단Method and device for generating matrix index information, and method and device for processing matrix using matrix index information

Similar Documents

PublicationPublication DateTitle
CN112418424A (en) A Hierarchical Sparse Coding Method for Pruned Deep Neural Networks with Extremely High Compression Ratio
US11403528B2 (en)Self-tuning incremental model compression solution in deep neural network with guaranteed accuracy performance
CN101183873B (en)BP neural network based embedded system data compression/decompression method
CN110796251A (en)Image compression optimization method based on convolutional neural network
CN108304928A (en)Compression method based on the deep neural network for improving cluster
CN111105007B (en) Compression and acceleration of deep convolutional neural networks for object detection
CN106557812A (en)The compression of depth convolutional neural networks and speeding scheme based on dct transform
CN110020721B (en) A deep learning network optimization method for target detection based on parameter compression
CN115022637B (en) Image encoding method, image decompression method and device
US12347150B2 (en)System and method for scene graph lossless compression by context-based graph convolution
CN116743182B (en)Lossless data compression method
CN111160524A (en)Two-stage convolutional neural network model compression method
CN117998086A (en)Lightweight image compression method and terminal
CN115361559A (en)Image encoding method, image decoding method, image encoding device, image decoding device, and storage medium
CN116961672A (en)Lossless data compression method based on transducer encoder
Zhe et al.Rate-distortion optimized coding for efficient cnn compression
CN116664705A (en)Heterogeneous calculation acceleration method for intelligent lossless compression algorithm
CN117692662A (en) A point cloud encoding and decoding method based on double octree structure
CN113902097B (en) Run-length encoding accelerator and method for sparse CNN neural network model
CN115471576A (en)Point cloud lossless compression method and device based on deep learning
CN115170613A (en)Human motion prediction method based on time sequence grading and recombination mechanism
CN114615507A (en)Image coding method, decoding method and related device
KR102394951B1 (en)Method and apparatus for encoding and decoding
CN109547961B (en) A Compressed Sensing Coding and Decoding Method for Large Data Volumes in Wireless Sensor Networks
CN111479286A (en) A data processing method for edge computing system to reduce communication traffic

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
RJ01Rejection of invention patent application after publication
RJ01Rejection of invention patent application after publication

Application publication date:20210226


[8]ページ先頭

©2009-2025 Movatter.jp