GitOrigin-RevId: 707cb88341
release-1.7
@@ -44,7 +44,6 @@ LITE_DYN_TYPE_OBJ_FINAL_IMPL(NetworkImplDft); | |||||
void NetworkImplDft::set_config(const Config& config) { | void NetworkImplDft::set_config(const Config& config) { | ||||
m_user_config = std::make_unique<Config>(); | m_user_config = std::make_unique<Config>(); | ||||
*m_user_config = config; | *m_user_config = config; | ||||
m_load_config.comp_graph = mgb::ComputingGraph::make(); | |||||
m_compnode_locator = to_compnode_locator(m_user_config->device_type); | m_compnode_locator = to_compnode_locator(m_user_config->device_type); | ||||
m_compnode_locator.device = config.device_id; | m_compnode_locator.device = config.device_id; | ||||
} | } | ||||
@@ -39,6 +39,7 @@ class NetworkImplDft final : public Network::NetworkImplBase { | |||||
LITE_DYN_TYPE_OBJ_FINAL_DECL; | LITE_DYN_TYPE_OBJ_FINAL_DECL; | ||||
public: | public: | ||||
NetworkImplDft() { m_load_config.comp_graph = mgb::ComputingGraph::make(); } | |||||
using S = megdnn::param::ExecutionPolicy::Strategy; | using S = megdnn::param::ExecutionPolicy::Strategy; | ||||
//! set the config of the network, include: | //! set the config of the network, include: | ||||
//! the inference device | //! the inference device | ||||
@@ -73,14 +73,14 @@ __attribute__((unused)) static std::shared_ptr<Tensor> get_input_data( | |||||
std::vector<npy::ndarray_len_t> stl_shape; | std::vector<npy::ndarray_len_t> stl_shape; | ||||
std::vector<int8_t> raw; | std::vector<int8_t> raw; | ||||
npy::LoadArrayFromNumpy(path, type_str, stl_shape, raw); | npy::LoadArrayFromNumpy(path, type_str, stl_shape, raw); | ||||
auto lite_tensor = std::make_shared<Tensor>(LiteDeviceType::LITE_CPU); | auto lite_tensor = std::make_shared<Tensor>(LiteDeviceType::LITE_CPU); | ||||
Layout layout; | Layout layout; | ||||
layout.ndim = stl_shape.size(); | layout.ndim = stl_shape.size(); | ||||
const std::map<std::string, LiteDataType> type_map = { | const std::map<std::string, LiteDataType> type_map = { | ||||
{"f4", LiteDataType::LITE_FLOAT}, | |||||
{"i4", LiteDataType::LITE_INT}, | |||||
{"i1", LiteDataType::LITE_INT8}, | |||||
{"f4", LiteDataType::LITE_FLOAT}, {"f2", LiteDataType::LITE_HALF}, | |||||
{"i8", LiteDataType::LITE_INT64}, {"i4", LiteDataType::LITE_INT}, | |||||
{"u4", LiteDataType::LITE_UINT}, {"i2", LiteDataType::LITE_INT16}, | |||||
{"u2", LiteDataType::LITE_UINT16}, {"i1", LiteDataType::LITE_INT8}, | |||||
{"u1", LiteDataType::LITE_UINT8}}; | {"u1", LiteDataType::LITE_UINT8}}; | ||||
layout.shapes[0] = 1; | layout.shapes[0] = 1; | ||||
for (size_t i = 0; i < stl_shape.size(); i++) { | for (size_t i = 0; i < stl_shape.size(); i++) { | ||||
@@ -109,7 +109,7 @@ def main(): | |||||
PackModel.PackModelStartModelsVector(builder, 1) | PackModel.PackModelStartModelsVector(builder, 1) | ||||
builder.PrependUOffsetTRelative(model) | builder.PrependUOffsetTRelative(model) | ||||
models = builder.EndVector(1) | |||||
models = builder.EndVector() | |||||
PackModel.PackModelStart(builder) | PackModel.PackModelStart(builder) | ||||
PackModel.PackModelAddModels(builder, models) | PackModel.PackModelAddModels(builder, models) | ||||