MNN/source/backend/vulkan/execution/VulkanROIPooling.cpp

89 lines
3.6 KiB
C++

//
// VulkanROIPooling.cpp
// MNN
//
// Created by MNN on 2019/01/31.
// Copyright © 2018, Alibaba Group Holding Limited
//
#include "backend/vulkan/execution/VulkanROIPooling.hpp"
#include "core/Macro.h"
namespace MNN {
struct GpuParam {
ivec4 inputImgSize;
ivec4 outputImgSize;
float spatialScale;
};
VulkanROIPooling::VulkanROIPooling(Backend* bn, const float SpatialScale)
: VulkanBasicExecution(bn), mSpatialScale(SpatialScale) {
std::vector<VkDescriptorType> VulkanROIPoolingTypes{
VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER,
};
auto extra = static_cast<VulkanBackend*>(bn);
mVulkanROIPoolingPipeline = extra->getPipeline(
"glsl_roipooling_comp",
/*"glsl_VulkanROIPooling_comp", glsl_roipooling_comp, glsl_roipooling_comp_len,*/ VulkanROIPoolingTypes);
mParamBuffer.reset(
new VulkanBuffer(extra->getMemoryPool(), false, sizeof(GpuParam), nullptr, VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT));
mSampler = extra->getCommonSampler();
}
VulkanROIPooling::~VulkanROIPooling() {
}
ErrorCode VulkanROIPooling::onEncode(const std::vector<Tensor*>& inputs, const std::vector<Tensor*>& outputs,
const VulkanCommandPool::Buffer* cmdBuffer) {
auto input = inputs[0];
auto roi = inputs[1];
auto output = outputs[0];
const int channelDiv4 = UP_DIV(input->channel(), 4);
auto VulkanROIPoolingParam = reinterpret_cast<GpuParam*>(mParamBuffer->map());
::memset(VulkanROIPoolingParam, 0, sizeof(GpuParam));
VulkanROIPoolingParam->inputImgSize[0] = input->width();
VulkanROIPoolingParam->inputImgSize[1] = input->height();
VulkanROIPoolingParam->inputImgSize[2] = channelDiv4;
VulkanROIPoolingParam->inputImgSize[3] = input->batch();
VulkanROIPoolingParam->outputImgSize[0] = output->width();
VulkanROIPoolingParam->outputImgSize[1] = output->height();
VulkanROIPoolingParam->outputImgSize[2] = channelDiv4;
VulkanROIPoolingParam->outputImgSize[3] = output->batch();
VulkanROIPoolingParam->spatialScale = mSpatialScale;
mParamBuffer->flush(true, 0, sizeof(GpuParam));
mParamBuffer->unmap();
mDescriptorSet.reset(mVulkanROIPoolingPipeline->createSet());
mDescriptorSet->writeImage(reinterpret_cast<VkImageView>(output->deviceId()), mSampler->get(),
VK_IMAGE_LAYOUT_GENERAL, 0);
mDescriptorSet->writeImage(reinterpret_cast<VkImageView>(input->deviceId()), mSampler->get(),
VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL, 1);
mDescriptorSet->writeImage(reinterpret_cast<VkImageView>(roi->deviceId()), mSampler->get(),
VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL, 2);
mDescriptorSet->writeBuffer(mParamBuffer->buffer(), 3, mParamBuffer->size());
mVulkanROIPoolingPipeline->bind(cmdBuffer->get(), mDescriptorSet->get());
vkCmdDispatch(cmdBuffer->get(), UP_DIV(output->width(), 8), UP_DIV(output->height(), 8),
channelDiv4 * output->batch());
return NO_ERROR;
}
class VulkanROIPoolingCreator : public VulkanBackend::Creator {
public:
virtual VulkanBasicExecution* onCreate(const std::vector<Tensor*>& inputs, const std::vector<Tensor*>& outputs, const MNN::Op* op, Backend* bn) const override {
return new VulkanROIPooling(bn, op->main_as_RoiPooling()->spatialScale());
}
};
static bool gResistor = []() {
VulkanBackend::addCreator(OpType_ROIPooling, new VulkanROIPoolingCreator);
return true;
}();
} // namespace MNN