mirror of https://github.com/alibaba/MNN.git
144 lines
5.9 KiB
C++
144 lines
5.9 KiB
C++
//
|
|
// OpenCLRunningUtils.hpp
|
|
// MNN
|
|
//
|
|
// Created by MNN on 2019/01/31.
|
|
// Copyright © 2018, Alibaba Group Holding Limited
|
|
//
|
|
|
|
#ifndef OpenCLRunningUtils_hpp
|
|
#define OpenCLRunningUtils_hpp
|
|
|
|
#include <string>
|
|
#include <vector>
|
|
#include <algorithm>
|
|
#include <climits>
|
|
|
|
#include "core/Macro.h"
|
|
#include "core/TensorUtils.hpp"
|
|
#include "backend/opencl/core/runtime/OpenCLRuntime.hpp"
|
|
#include "backend/opencl/core/runtime/OpenCLWrapper.hpp"
|
|
#include "backend/opencl/core/BufferPool.hpp"
|
|
|
|
namespace MNN {
|
|
namespace OpenCL {
|
|
|
|
enum CLTuneLevel { None = 0, Heavy = 1, Wide = 2, Normal = 3, Fast = 4};
|
|
enum GpuMemObject { AUTO = 0, BUFFER = 1, IMAGE = 2};
|
|
inline std::vector<int> tensorShapeFormat(const Tensor *input) {
|
|
|
|
int iN = (0 != input->buffer().dim[0].extent) ? input->buffer().dim[0].extent : 1;
|
|
int iC = (0 != input->buffer().dim[1].extent) ? input->buffer().dim[1].extent : 1;
|
|
int iH = (0 != input->buffer().dim[2].extent) ? input->buffer().dim[2].extent : 1;
|
|
int iW = (0 != input->buffer().dim[3].extent) ? input->buffer().dim[3].extent : 1;
|
|
|
|
if(input->buffer().dimensions > 4)//more than 4 dimensions put to N dimension
|
|
{
|
|
for(int i = 4; i < input->buffer().dimensions; i++)
|
|
{
|
|
iW *= input->buffer().dim[i].extent;
|
|
}
|
|
}
|
|
|
|
if (TensorUtils::getDescribe(input)->dimensionFormat == MNN::MNN_DATA_FORMAT_NHWC)
|
|
{
|
|
iN = (0 < input->buffer().dim[0].extent) ? input->buffer().dim[0].extent : 1;
|
|
iH = (0 < input->buffer().dim[1].extent) ? input->buffer().dim[1].extent : 1;
|
|
iW = (0 < input->buffer().dim[2].extent) ? input->buffer().dim[2].extent : 1;
|
|
iC = (0 < input->buffer().dim[3].extent) ? input->buffer().dim[3].extent : 1;
|
|
|
|
if(input->buffer().dimensions > 4)//more than 4 dimensions put to N dimension
|
|
{
|
|
for(int i = 4; i < input->buffer().dimensions; i++)
|
|
{
|
|
iC *= input->buffer().dim[i].extent;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (input->buffer().dimensions == 2) {
|
|
iN = input->buffer().dim[0].extent;
|
|
iH = 1;
|
|
iW = 1;
|
|
iC = input->buffer().dim[1].extent;
|
|
}
|
|
if (input->buffer().dimensions == 1) {
|
|
iN = 1;
|
|
iH = 1;
|
|
iW = 1;
|
|
iC = input->buffer().dim[0].extent;
|
|
}
|
|
|
|
#ifdef LOG_VERBOSE
|
|
MNN_PRINT("tensorShapeFormat : [%d, %d, %d, %d] \n", iN, iH, iW, iC);
|
|
#endif
|
|
std::vector<int> shape_vec{iN, iH, iW, iC};
|
|
|
|
return shape_vec;
|
|
}
|
|
|
|
enum OpenCLBufferFormat {
|
|
CONV2D_FILTER = 0,
|
|
NHWC_BUFFER = 1,
|
|
ARGUMENT = 2,
|
|
DW_CONV2D_FILTER = 3,
|
|
NCHW_BUFFER = 4,
|
|
NHWC4_BUFFER = 5,
|
|
CONV2D1x1_OPT_FILTER = 6,
|
|
};
|
|
|
|
template <typename T, typename Dim>
|
|
inline void IOHW2OIHW(const T *src, T *dst, Dim O, Dim I, Dim H, Dim W) {
|
|
for (Dim i = 0; i < I; i++) {
|
|
for (Dim o = 0; o < O; o++) {
|
|
for (Dim h = 0; h < H; h++) {
|
|
for (Dim w = 0; w < W; w++) {
|
|
dst[o * I * H * W + i * H * W + h * W + w] = src[i * O * H * W + o * H * W + h * W + w];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
};
|
|
inline cl::Buffer &openCLDeferBuffer(const Tensor *tensor) {
|
|
return *(*(OpenCLBufferNode *)(tensor->deviceId())).buffer.get();
|
|
}
|
|
inline cl::Buffer &openCLBuffer(const Tensor *tensor) {
|
|
return (*(cl::Buffer *)(tensor->deviceId()));
|
|
}
|
|
inline cl::Image &openCLImage(const Tensor *tensor) {
|
|
return (*(cl::Image *)(tensor->deviceId()));
|
|
}
|
|
|
|
void getImageShape(const std::vector<int> &shape, /* NHWC */
|
|
const OpenCLBufferFormat type, std::vector<size_t> *imageShape);
|
|
|
|
void run3DKernelDefault(const ::std::shared_ptr<KernelWrap> &kernel, const std::vector<uint32_t> &gws, const std::vector<uint32_t> &lws,
|
|
OpenCLRuntime *runtime, cl::Event* eventPtr = nullptr);
|
|
|
|
void runKernel2D(const ::std::shared_ptr<KernelWrap> &kernel, const std::vector<uint32_t> &gws, const std::vector<uint32_t> &lws,
|
|
OpenCLRuntime *runtime, cl::Event* eventPtr = nullptr);
|
|
|
|
void runTurnKernelLWS2D(const ::std::shared_ptr<KernelWrap> &kernel, const std::vector<uint32_t> &gws, const std::vector<uint32_t> &lws,
|
|
OpenCLRuntime *runtime);
|
|
std::vector<uint32_t> getGemmParams(const std::vector<uint32_t> &gemmSize, const std::vector<cl::Buffer> tensorMemory,
|
|
OpenCLRuntime *runtime, int precision, int tuneLevel);
|
|
std::pair<std::vector<uint32_t>, uint32_t> localWS3DDefault(const std::vector<uint32_t> &gws, const uint32_t maxWorkGroupSize,
|
|
OpenCLRuntime *runtime, const std::string &kernelName, const std::shared_ptr<KernelWrap> &mKernel, int tuneLevel);
|
|
|
|
bool localWSTune(const std::map<std::string, std::vector<std::pair<std::vector<uint32_t>, std::pair<std::vector<uint32_t>, uint32_t>>>> &tuneMap, const std::vector<uint32_t> &gws, const std::string &kernelName, std::pair<std::vector<uint32_t>, uint32_t> &res);
|
|
|
|
uint32_t get2DUseLocalMemTime(const std::vector<uint32_t> &gws, const std::vector<uint32_t> &lws, OpenCLRuntime *runtime, const std::string &kernelName, const std::shared_ptr<KernelWrap> &mKernelW);
|
|
|
|
std::pair<std::vector<uint32_t>, uint32_t> localWS2DDefault(const std::vector<uint32_t> &gws, const uint32_t maxWorkGroupSize,
|
|
OpenCLRuntime *runtime, const std::string &kernelName, const std::shared_ptr<KernelWrap> &mKernel, int tuneLevel);
|
|
|
|
bool getTunedInfo(const std::string kernelName, const std::vector<uint32_t> &gws, std::pair<std::vector<uint32_t>, uint32_t> &tuneInfo, OpenCLRuntime *runtime);
|
|
|
|
void setTunedInfo(const std::string kernelName, const std::vector<uint32_t> &gws, std::pair<std::vector<uint32_t>, uint32_t> &tuneInfo, OpenCLRuntime *runtime);
|
|
|
|
void copyBufferToImage(OpenCLRuntime *runtime, const cl::Buffer &buffer, const cl::Image &image, int w, int h, int precision);
|
|
|
|
} // namespace OpenCL
|
|
} // namespace MNN
|
|
#endif /* OpenCLRunningUtils_hpp */
|