mirror of https://github.com/alibaba/MNN.git
				
				
				
			
		
			
				
	
	
		
			481 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			481 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			C++
		
	
	
	
//
 | 
						|
//  Interpreter.cpp
 | 
						|
//  MNN
 | 
						|
//
 | 
						|
//  Created by MNN on 2018/07/30.
 | 
						|
//  Copyright © 2018, Alibaba Group Holding Limited
 | 
						|
//
 | 
						|
 | 
						|
#include <math.h>
 | 
						|
#include <stdio.h>
 | 
						|
#include <MNN/Interpreter.hpp>
 | 
						|
#include <algorithm>
 | 
						|
#include <mutex>
 | 
						|
#include <vector>
 | 
						|
#include "MNN_generated.h"
 | 
						|
#include "core/AutoStorage.h"
 | 
						|
#include "core/FileLoader.hpp"
 | 
						|
#include "core/Pipeline.hpp"
 | 
						|
#include "core/RuntimeFactory.hpp"
 | 
						|
#include "core/Session.hpp"
 | 
						|
 | 
						|
namespace MNN {
 | 
						|
 | 
						|
struct Content {
 | 
						|
    AutoStorage<uint8_t> buffer;
 | 
						|
    const Net* net = nullptr;
 | 
						|
    std::vector<std::unique_ptr<Session>> sessions;
 | 
						|
    std::map<const Tensor*, const Session*> tensorMap;
 | 
						|
    Interpreter::SessionMode callBackMode = Interpreter::Session_Debug;
 | 
						|
    Interpreter::SessionMode inputMode    = Interpreter::Session_Input_Inside;
 | 
						|
    AutoStorage<uint8_t> cacheBuffer;
 | 
						|
    size_t cacheOffset = 0;
 | 
						|
    std::string cacheFile;
 | 
						|
    std::mutex lock;
 | 
						|
    size_t lastCacheSize = 0;
 | 
						|
};
 | 
						|
 | 
						|
static void writeCacheFile(const Content *net, std::pair<const void*, size_t> buffer) {
 | 
						|
    FILE* f = fopen(net->cacheFile.c_str(), "wb");
 | 
						|
    if (nullptr == f) {
 | 
						|
        MNN_ERROR("Open %s error\n", net->cacheFile.c_str());
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    // Write key
 | 
						|
    auto tsize = fwrite((const char*)net->buffer.get(), 1, net->cacheOffset, f);
 | 
						|
    if (tsize != net->cacheOffset) {
 | 
						|
        MNN_ERROR("Write %s error\n", net->cacheFile.c_str());
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    // Write Cache
 | 
						|
    static const size_t block = 4096;
 | 
						|
    size_t totalSize          = buffer.second;
 | 
						|
    size_t blockSize          = UP_DIV(totalSize, block);
 | 
						|
    for (size_t i = 0; i < blockSize; ++i) {
 | 
						|
        size_t sta = block * i;
 | 
						|
        size_t fin = std::min(sta + block, totalSize);
 | 
						|
        if (fin > sta) {
 | 
						|
            auto realSize = fwrite((const char*)(buffer.first) + sta, 1, fin - sta, f);
 | 
						|
            if (realSize != fin - sta) {
 | 
						|
                MNN_ERROR("Write %s error\n", net->cacheFile.c_str());
 | 
						|
                return;
 | 
						|
            }
 | 
						|
        }
 | 
						|
    }
 | 
						|
    fclose(f);
 | 
						|
}
 | 
						|
 | 
						|
Interpreter* Interpreter::createFromFile(const char* file) {
 | 
						|
    if (nullptr == file) {
 | 
						|
        MNN_PRINT("NULL file for create interpreter\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    std::unique_ptr<FileLoader> loader(new FileLoader(file));
 | 
						|
    if (!loader->valid()) {
 | 
						|
        MNN_PRINT("Create interpreter failed, open %s error\n", file);
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    bool result = loader->read();
 | 
						|
    if (!result) {
 | 
						|
        MNN_PRINT("Read file error\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    if (loader->size() == 0) {
 | 
						|
        MNN_PRINT("Create interpreter failed, %s is empty\n", file);
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    auto net     = new Content;
 | 
						|
    bool success = loader->merge(net->buffer);
 | 
						|
    if (!success) {
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    loader.reset();
 | 
						|
    return createFromBufferInternal(net);
 | 
						|
}
 | 
						|
Interpreter* Interpreter::createFromBuffer(const void* buffer, size_t size) {
 | 
						|
    if (nullptr == buffer || 0 == size) {
 | 
						|
        MNN_PRINT("Buffer is null for create interpreter\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    auto net = new Content;
 | 
						|
    net->buffer.reset((int)size);
 | 
						|
    if (nullptr == net->buffer.get()) {
 | 
						|
        MNN_ERROR("Memory not enought!\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    ::memcpy(net->buffer.get(), buffer, size);
 | 
						|
 | 
						|
    return createFromBufferInternal(net);
 | 
						|
}
 | 
						|
 | 
						|
Interpreter* Interpreter::createFromBufferInternal(Content* net) {
 | 
						|
    if (nullptr == net) {
 | 
						|
        MNN_PRINT("Buffer is null for create interpreter\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
#ifndef MNN_BUILD_MINI
 | 
						|
    flatbuffers::Verifier verify((const uint8_t*)(net->buffer.get()), net->buffer.size());
 | 
						|
    if (false == VerifyNetBuffer(verify)) {
 | 
						|
        MNN_PRINT("Invalidate buffer to create interpreter\n");
 | 
						|
        delete net;
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
#endif
 | 
						|
    net->net = GetNet(net->buffer.get());
 | 
						|
    if (nullptr == net->net->oplists()) {
 | 
						|
        MNN_ERROR("Model has no oplist\n");
 | 
						|
        delete net;
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    int opSize = net->net->oplists()->size();
 | 
						|
    for (int i = 0; i < opSize; ++i) {
 | 
						|
        auto op = net->net->oplists()->GetAs<Op>(i);
 | 
						|
        if (nullptr == op || nullptr == op->outputIndexes()) {
 | 
						|
            MNN_ERROR("Invalid Model, the %d op is empty\n", i);
 | 
						|
            delete net;
 | 
						|
            return nullptr;
 | 
						|
        }
 | 
						|
    }
 | 
						|
    return new Interpreter(net);
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::setSessionMode(SessionMode mode) {
 | 
						|
    if (mode == Session_Input_Inside || mode == Session_Input_User) {
 | 
						|
        mNet->inputMode = mode;
 | 
						|
    } else {
 | 
						|
        mNet->callBackMode = mode;
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::setCacheFile(const char* cacheFile, size_t keySize) {
 | 
						|
    if (nullptr == cacheFile || nullptr == mNet->buffer.get()) {
 | 
						|
        MNN_ERROR("Empty cacheFile or the interpreter invalid\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    mNet->cacheFile   = std::string(cacheFile);
 | 
						|
    mNet->cacheOffset = mNet->buffer.size() > keySize ? keySize : mNet->buffer.size();
 | 
						|
    std::unique_ptr<FileLoader> loader(new FileLoader(cacheFile));
 | 
						|
    if (!loader->valid()) {
 | 
						|
        MNN_ERROR("Load Cache file error.\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    bool result = loader->read();
 | 
						|
    if (!result) {
 | 
						|
        MNN_ERROR("Load Cache file error.\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    if (loader->size() == 0) {
 | 
						|
        MNN_ERROR("Load Cache file error.\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    bool success = loader->merge(mNet->cacheBuffer);
 | 
						|
    if (!success) {
 | 
						|
        MNN_ERROR("Alloc memory for Cache error.\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
ErrorCode Interpreter::updateCacheFile(Session *session, int flag) {
 | 
						|
    auto buffer = session->getCache();
 | 
						|
    
 | 
						|
    //When current cacheSize bigger than previous, update
 | 
						|
    if (buffer.first != nullptr && buffer.second > mNet->lastCacheSize) {
 | 
						|
        MNN_PRINT("Update cache to %s, size = %zu\n", mNet->cacheFile.c_str(), buffer.second);
 | 
						|
        writeCacheFile(mNet, buffer);
 | 
						|
        mNet->lastCacheSize = buffer.second;
 | 
						|
    }
 | 
						|
    // Reset cache
 | 
						|
    session->loadCache(nullptr, 0);
 | 
						|
    return NO_ERROR;
 | 
						|
}
 | 
						|
 | 
						|
Interpreter::Interpreter(Content* net) {
 | 
						|
    MNN_ASSERT(nullptr != net);
 | 
						|
    mNet = net;
 | 
						|
}
 | 
						|
 | 
						|
Interpreter::~Interpreter() {
 | 
						|
    {
 | 
						|
        // If the session is running, we must not delete session
 | 
						|
        std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
        mNet->sessions.clear();
 | 
						|
        mNet->tensorMap.clear();
 | 
						|
    }
 | 
						|
    delete mNet;
 | 
						|
}
 | 
						|
 | 
						|
Session* Interpreter::createMultiPathSession(const std::vector<ScheduleConfig>& configs) {
 | 
						|
    RuntimeInfo runtime = createRuntime(configs);
 | 
						|
    if (runtime.first.empty()) {
 | 
						|
        MNN_ERROR("Runtime not valid for create session\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    return createMultiPathSession(configs, std::move(runtime));
 | 
						|
}
 | 
						|
 | 
						|
Session* Interpreter::createMultiPathSession(const std::vector<ScheduleConfig>& configs, const RuntimeInfo& runtime) {
 | 
						|
    if (nullptr == mNet->buffer.get()) {
 | 
						|
        MNN_ERROR("The model buffer has been released. Can't create session\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    if (runtime.first.empty()) {
 | 
						|
        MNN_ERROR("Runtime not valid for create session\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    bool netBufferHold  = mNet->inputMode == Session_Input_User;
 | 
						|
    Schedule::ScheduleInfo info;
 | 
						|
    auto success = Schedule::schedule(info, mNet->net, configs, runtime, netBufferHold);
 | 
						|
    if (!success) {
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    auto validForResize = info.validForResize;
 | 
						|
    RuntimeInfo rt = runtime;
 | 
						|
    auto newSession =
 | 
						|
        std::unique_ptr<Session>(new Session(std::move(info), mNet->callBackMode, mNet->inputMode, std::move(rt)));
 | 
						|
    if (!newSession->valid()) {
 | 
						|
        MNN_PRINT("Invalide Session!!\n");
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    auto result = newSession.get();
 | 
						|
    bool valid  = false;
 | 
						|
    if (mNet->cacheBuffer.get() != nullptr) {
 | 
						|
        valid = result->loadCache(mNet->cacheBuffer.get() + mNet->cacheOffset,
 | 
						|
                                  mNet->cacheBuffer.size() - mNet->cacheOffset);
 | 
						|
        if(!valid) {
 | 
						|
            // Reset cache
 | 
						|
            result->loadCache(nullptr, 0);
 | 
						|
            MNN_PRINT("Cache invalid, will be reset\n");
 | 
						|
        }
 | 
						|
        
 | 
						|
        mNet->lastCacheSize = mNet->cacheBuffer.size() - mNet->cacheOffset;
 | 
						|
    }
 | 
						|
    if (validForResize && mNet->inputMode == Session_Input_Inside) {
 | 
						|
        result->resize(mNet->net->usage() == Usage_INFERENCE_STATIC);
 | 
						|
    }
 | 
						|
    
 | 
						|
    if ((!mNet->cacheFile.empty()) && (!valid)) {
 | 
						|
        // Try to save extra cache
 | 
						|
        auto buffer = result->getCache();
 | 
						|
        if (buffer.first != nullptr && buffer.second > 0) {
 | 
						|
            MNN_PRINT("Write cache to %s, size = %zu\n", mNet->cacheFile.c_str(), buffer.second);
 | 
						|
            writeCacheFile(mNet, buffer);
 | 
						|
            mNet->lastCacheSize = buffer.second;
 | 
						|
        }
 | 
						|
    }
 | 
						|
    // Reset cache
 | 
						|
    result->loadCache(nullptr, 0);
 | 
						|
 | 
						|
    mNet->sessions.emplace_back(std::move(newSession));
 | 
						|
    
 | 
						|
    //for valid cacheFile, maybe need update after resize
 | 
						|
    if(valid) {
 | 
						|
        updateCacheFile(result);
 | 
						|
    }
 | 
						|
    return result;
 | 
						|
}
 | 
						|
 | 
						|
Session* Interpreter::createSession(const ScheduleConfig& config) {
 | 
						|
    return createMultiPathSession({config});
 | 
						|
}
 | 
						|
 | 
						|
Session* Interpreter::createSession(const ScheduleConfig& config, const RuntimeInfo& runtime) {
 | 
						|
    return createMultiPathSession({config}, runtime);
 | 
						|
}
 | 
						|
 | 
						|
bool Interpreter::releaseSession(Session* session) {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    for (auto iter = mNet->sessions.begin(); iter != mNet->sessions.end(); iter++) {
 | 
						|
        // TODO Delete tensormap
 | 
						|
        for (auto tIter = mNet->tensorMap.begin(); tIter != mNet->tensorMap.end();) {
 | 
						|
            if (tIter->second == session) {
 | 
						|
                tIter = mNet->tensorMap.erase(tIter);
 | 
						|
                continue;
 | 
						|
            }
 | 
						|
            tIter++;
 | 
						|
        }
 | 
						|
 | 
						|
        if ((*iter).get() == session) {
 | 
						|
            mNet->sessions.erase(iter);
 | 
						|
            return true;
 | 
						|
        }
 | 
						|
    }
 | 
						|
    return false;
 | 
						|
}
 | 
						|
 | 
						|
ErrorCode Interpreter::runSession(Session* session) const {
 | 
						|
    return session->run();
 | 
						|
}
 | 
						|
 | 
						|
Tensor* Interpreter::getSessionInput(const Session* session, const char* name) {
 | 
						|
    if (session == nullptr) {
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    auto tensor = session->getInput(name);
 | 
						|
    mNet->tensorMap.insert(std::make_pair(tensor, session));
 | 
						|
    return tensor;
 | 
						|
}
 | 
						|
 | 
						|
Tensor* Interpreter::getSessionOutput(const Session* session, const char* name) {
 | 
						|
    if (session == nullptr) {
 | 
						|
        return nullptr;
 | 
						|
    }
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    auto tensor = session->getOutput(name);
 | 
						|
    mNet->tensorMap.insert(std::make_pair(tensor, session));
 | 
						|
    return tensor;
 | 
						|
}
 | 
						|
 | 
						|
const std::map<std::string, Tensor*>& Interpreter::getSessionInputAll(const Session* session) const {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    auto& tensors = session->getInputAll();
 | 
						|
    for (auto& iter : tensors) {
 | 
						|
        mNet->tensorMap.insert(std::make_pair(iter.second, session));
 | 
						|
    }
 | 
						|
    return tensors;
 | 
						|
}
 | 
						|
 | 
						|
const std::map<std::string, Tensor*>& Interpreter::getSessionOutputAll(const Session* session) const {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    auto& tensors = session->getOutputAll();
 | 
						|
    for (auto& iter : tensors) {
 | 
						|
        mNet->tensorMap.insert(std::make_pair(iter.second, session));
 | 
						|
    }
 | 
						|
    return tensors;
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::resizeSession(Session* session) {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    if (mNet->buffer.get() == nullptr) {
 | 
						|
        MNN_ERROR("The model buffer has been released. Can't resize session\n");
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    session->resize();
 | 
						|
}
 | 
						|
 | 
						|
ErrorCode Interpreter::runSessionWithCallBack(const Session* session, const TensorCallBack& before,
 | 
						|
                                              const TensorCallBack& after, bool sync) const {
 | 
						|
    auto beforeWrap = [&before](const std::vector<Tensor*>& tensors, const OperatorInfo* info) {
 | 
						|
        return before(tensors, info->name());
 | 
						|
    };
 | 
						|
    auto afterWrap = [&after](const std::vector<Tensor*>& tensors, const OperatorInfo* info) {
 | 
						|
        return after(tensors, info->name());
 | 
						|
    };
 | 
						|
    return runSessionWithCallBackInfo(session, beforeWrap, afterWrap, sync);
 | 
						|
}
 | 
						|
 | 
						|
ErrorCode Interpreter::runSessionWithCallBackInfo(const Session* session, const TensorCallBackWithInfo& before,
 | 
						|
                                                  const TensorCallBackWithInfo& callBack, bool sync) const {
 | 
						|
    return session->runWithCallBack(before, callBack, sync);
 | 
						|
}
 | 
						|
 | 
						|
const Backend* Interpreter::getBackend(const Session* session, const Tensor* tensor) const {
 | 
						|
    return session->getBackEnd(tensor);
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::releaseModel() {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    if (mNet->buffer.get() != nullptr && mNet->net->usage() != Usage_INFERENCE_STATIC) {
 | 
						|
        mNet->buffer.release();
 | 
						|
    }
 | 
						|
    mNet->cacheBuffer.release();
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::resizeTensor(Tensor* tensor, int batch, int channel, int height, int width) {
 | 
						|
    if (tensor->getDimensionType() == Tensor::TENSORFLOW) {
 | 
						|
        resizeTensor(tensor, {batch, height, width, channel});
 | 
						|
    } else {
 | 
						|
        resizeTensor(tensor, {batch, channel, height, width});
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
void Interpreter::resizeTensor(Tensor* tensor, const std::vector<int>& dims) {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    MNN_ASSERT(nullptr != tensor);
 | 
						|
    bool dirty = false;
 | 
						|
    if (tensor->buffer().dimensions != dims.size()) {
 | 
						|
        dirty = true;
 | 
						|
    } else {
 | 
						|
        for (int i = 0; i < dims.size(); ++i) {
 | 
						|
            if (tensor->buffer().dim[i].extent != dims[i]) {
 | 
						|
                dirty = true;
 | 
						|
                break;
 | 
						|
            }
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    if (!dirty) {
 | 
						|
        return;
 | 
						|
    }
 | 
						|
 | 
						|
    tensor->buffer().dimensions = (int)dims.size();
 | 
						|
    for (int i = 0; i < dims.size(); ++i) {
 | 
						|
        tensor->buffer().dim[i].extent = dims[i];
 | 
						|
    }
 | 
						|
 | 
						|
    auto relatedSessionIter = mNet->tensorMap.find(tensor);
 | 
						|
    MNN_ASSERT(relatedSessionIter != mNet->tensorMap.end());
 | 
						|
    ((MNN::Session*)relatedSessionIter->second)->setNeedResize();
 | 
						|
}
 | 
						|
 | 
						|
const char* Interpreter::bizCode() const {
 | 
						|
    const flatbuffers::String* code = mNet->net->bizCode();
 | 
						|
    return code->c_str();
 | 
						|
}
 | 
						|
 | 
						|
std::pair<const void*, size_t> Interpreter::getModelBuffer() const {
 | 
						|
    return std::make_pair(mNet->buffer.get(), mNet->buffer.size());
 | 
						|
}
 | 
						|
ErrorCode Interpreter::updateSessionToModel(Session* session) {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    if (mNet->buffer.get() == nullptr) {
 | 
						|
        MNN_ERROR("Can't updateSessionToModel because you called releaseModel before\n");
 | 
						|
        return INPUT_DATA_ERROR;
 | 
						|
    }
 | 
						|
    return session->updateToModel((Net*)mNet->net);
 | 
						|
}
 | 
						|
 | 
						|
bool Interpreter::getSessionInfo(const Session* session, SessionInfoCode code, void* ptr) {
 | 
						|
    std::unique_lock<std::mutex> _l(mNet->lock);
 | 
						|
    if (nullptr == session || nullptr == ptr) {
 | 
						|
        return false;
 | 
						|
    }
 | 
						|
    return session->getInfo(code, ptr);
 | 
						|
}
 | 
						|
 | 
						|
static void _getDefaultBackend(RuntimeInfo& rt) {
 | 
						|
    auto defaultType = MNN_FORWARD_CPU;
 | 
						|
    if (rt.first.find(defaultType) != rt.first.end()) {
 | 
						|
        rt.second = rt.first[defaultType];
 | 
						|
    }
 | 
						|
    if (rt.second == nullptr) {
 | 
						|
        Backend::Info info;
 | 
						|
        info.type      = defaultType;
 | 
						|
        info.numThread = 1;
 | 
						|
        rt.second.reset(RuntimeFactory::create(info));
 | 
						|
    }
 | 
						|
}
 | 
						|
RuntimeInfo Interpreter::createRuntime(const std::vector<ScheduleConfig>& configs) {
 | 
						|
    RuntimeInfo res;
 | 
						|
    auto& mRuntimes = res.first;
 | 
						|
    for (auto& config : configs) {
 | 
						|
        Backend::Info compute;
 | 
						|
        compute.type      = Schedule::getApprociateType(config);
 | 
						|
        compute.numThread = config.numThread;
 | 
						|
        compute.user      = config.backendConfig;
 | 
						|
        if (mRuntimes.find(compute.type) == mRuntimes.end()) {
 | 
						|
            auto newBn = RuntimeFactory::create(compute);
 | 
						|
            if (nullptr == newBn) {
 | 
						|
                MNN_ERROR("Can't create Runtime: %s\n", EnumNameForwardType((ForwardType)compute.type));
 | 
						|
                continue;
 | 
						|
            }
 | 
						|
            mRuntimes[compute.type].reset(newBn);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    _getDefaultBackend(res);
 | 
						|
    return res;
 | 
						|
}
 | 
						|
 | 
						|
} // namespace MNN
 |