MNN/source/backend/cpu/compute/SparseConvolutionTiledExecu...

65 lines
3.0 KiB
C++
Raw Normal View History

//
// SparseConvolutionTiledExecutor
// MNN
//
// Created by MNN on 2021/04/06.
// Copyright © 2018-2021 Alibaba Group Holding Limited.
//
#ifndef SparseConvolutionTiledExecutor_hpp
#define SparseConvolutionTiledExecutor_hpp
#include <functional>
#include "backend/cpu/CPUConvolution.hpp"
#include "ConvolutionTiledExecutor.hpp"
// Tiled Slide Window or Im2Col + GEMM
#define SPARSITY_THRESHOLD (0.3f)
namespace MNN {
class SparseConvolutionTiledImpl : public ConvolutionTiledImpl {
public:
SparseConvolutionTiledImpl(const Convolution2DCommon *common, const SparseCommon* sparseCommon, Backend *b) : mSparseCommon{sparseCommon}, ConvolutionTiledImpl(common, b) {
mSparseBlockOC = mSparseCommon->args()->LookupByKey("sparseBlockOC")->i();
}
virtual ~SparseConvolutionTiledImpl() = default;
virtual ErrorCode onResize(const std::vector<Tensor *> &inputs, const std::vector<Tensor *> &outputs, Tensor* NNZMap, Tensor* dataOffsetMap);
void getPackParameter(int* eP, int* lP, int* hP, const CoreFunctions* core) override;
public:
const SparseCommon* mSparseCommon;
int mSparseBlockOC;
};
class SparseConvolutionTiledExecutor : public ConvolutionTiledExecutor {
public:
SparseConvolutionTiledExecutor(const Convolution2DCommon *common, Backend *b, const float *originWeight,
size_t originWeightSize, const SparseCommon* sparseCommon, const float *bias, size_t biasSize);
SparseConvolutionTiledExecutor(std::shared_ptr<CPUConvolution::Resource> res, std::shared_ptr<Tensor> NNZMapSharePtr, std::shared_ptr<Tensor> dataOffsetMapSharePtr,
const Convolution2DCommon *common, const SparseCommon* sparseCommon, Backend *b);
virtual ~SparseConvolutionTiledExecutor();
virtual ErrorCode onExecute(const std::vector<Tensor *> &inputs, const std::vector<Tensor *> &outputs) override {
return mProxy->onExecute(inputs, outputs);
}
virtual ErrorCode onResize(const std::vector<Tensor *> &inputs, const std::vector<Tensor *> &outputs) override {
mInputs = {inputs[0], mResource->mWeight.get(), mResource->mBias.get()};
return mProxy->onResize(mInputs, outputs, mNNZMap.get(), mDataOffsetMap.get());
}
virtual bool onClone(Backend *bn, const Op *op, Execution **dst) override;
void initWeight(float *dest, unsigned int *NNZMap, int *dataOffsetMap, int sparseBlockOC, const float *source,
float *cache, int depth, int outputCount, int kernelSize, int eP, size_t weightNNZElement,
size_t weightBlockNumber, const CoreFunctions *function);
static bool shouldUseSparseConvolution(size_t originWeightSize, const SparseCommon* sparseCommon) {
return originWeightSize - sparseCommon->args()->LookupByKey("NNZElement")->i() >= originWeightSize * SPARSITY_THRESHOLD;
}
protected:
std::shared_ptr<SparseConvolutionTiledImpl> mProxy;
std::shared_ptr<Tensor> mNNZMap;
std::shared_ptr<Tensor> mDataOffsetMap;
};
} // namespace MNN
#endif /* SparseConvolutionTiledExecutor_hpp */