Browse Source

Merge branch 'development' of https://gitee.com/zhao_zhixuan/graphengine into development

tags/v1.1.0
unknown 4 years ago
parent
commit
5f152c6dbf
31 changed files with 114 additions and 200 deletions
  1. +2
    -0
      .gitmodules
  2. +13
    -14
      CMakeLists.txt
  3. +2
    -2
      cmake/external_libs/json.cmake
  4. +27
    -0
      ge/ge_runtime/proto/task.pb.h
  5. +29
    -3
      ge/generator/ge_generator.cc
  6. +14
    -2
      ge/graph/build/memory/block_mem_assigner.cc
  7. +1
    -1
      ge/graph/build/memory/block_mem_assigner.h
  8. +11
    -6
      ge/ir_build/ge_ir_build.cc
  9. +5
    -10
      ge/offline/single_op_parser.cc
  10. +8
    -8
      inc/framework/generator/ge_generator.h
  11. +1
    -1
      metadef
  12. +1
    -1
      parser
  13. +0
    -1
      tests/depends/cce/src/cce_stub.cc
  14. +0
    -2
      tests/st/resnet50/resnet50_train.cc
  15. +0
    -2
      tests/ut/common/graph/testcase/ge_graph/ge_model_unittest.cc
  16. +0
    -16
      tests/ut/ge/graph/load/new_model_manager_davinci_model_unittest.cc
  17. +0
    -95
      tests/ut/ge/graph/load/new_model_manager_model_manager_unittest.cc
  18. +0
    -5
      tests/ut/ge/graph/load/new_op_test_utils.h
  19. +0
    -1
      tests/ut/ge/graph/passes/dimension_adjust_pass_unittest.cc
  20. +0
    -15
      tests/ut/ge/graph/passes/folding_kernel/strided_slice_kernel_unittest.cc
  21. +0
    -1
      tests/ut/ge/graph/passes/guarantee_const_pass_unittest.cc
  22. +0
    -1
      tests/ut/ge/graph/passes/identity_pass_unittest.cc
  23. +0
    -1
      tests/ut/ge/graph/passes/net_output_pass_unittest.cc
  24. +0
    -1
      tests/ut/ge/graph/passes/placeholder_with_default_pass_unittest.cc
  25. +0
    -1
      tests/ut/ge/graph/passes/prevent_gradient_pass_unittest.cc
  26. +0
    -1
      tests/ut/ge/graph/passes/reshape_remove_pass_unittest.cc
  27. +0
    -1
      tests/ut/ge/graph/passes/snapshot_pass_unittest.cc
  28. +0
    -1
      tests/ut/ge/graph/passes/stop_gradient_pass_unittest.cc
  29. +0
    -2
      tests/ut/ge/graph/passes/switch_pass_unittest.cc
  30. +0
    -1
      tests/ut/ge/graph/passes/unused_and_isolated_op_remove_pass_unittest.cc
  31. +0
    -4
      tests/ut/ge/graph/passes/variable_op_pass_unittest.cc

+ 2
- 0
.gitmodules View File

@@ -1,6 +1,8 @@
[submodule "parser"]
path = parser
url = https://gitee.com/ascend/parser.git
branch = development
[submodule "metadef"]
path = metadef
url = https://gitee.com/ascend/metadef.git
branch = development

+ 13
- 14
CMakeLists.txt View File

@@ -62,12 +62,13 @@ if (ENABLE_OPEN_SRC)
endif()
set(ASCEND_DRIVER_DIR ${ASCEND_DIR}/driver/lib64)
set(ASCEND_DRIVER_COMMON_DIR ${ASCEND_DIR}/driver/lib64/common)
set(ASCEND_DRIVER_SHARE_DIR ${ASCEND_DIR}/driver/lib64/share)
set(ASCEND_RUNTIME_DIR ${ASCEND_DIR}/fwkacllib/lib64)
set(ASCEND_ATC_DIR ${ASCEND_DIR}/atc/lib64)
set(ASCEND_ACL_DIR ${ASCEND_DIR}/acllib/lib64)
find_module(slog libslog.so ${ASCEND_ATC_DIR})
find_module(mmpa libmmpa.so ${ASCEND_ATC_DIR})
if(PLATFORM STREQUAL "train")
find_module(slog libslog.so ${ASCEND_ATC_DIR})
find_module(mmpa libmmpa.so ${ASCEND_ATC_DIR})
find_module(msprof libmsprof.so ${ASCEND_DRIVER_COMMON_DIR})
find_module(hccl libhccl.so ${ASCEND_RUNTIME_DIR})
find_module(adump_server libadump_server.a ${ASCEND_RUNTIME_DIR})
@@ -80,8 +81,6 @@ if (ENABLE_OPEN_SRC)
message(FATAL_ERROR "This platform is not supported in train mode, build terminated")
endif()
elseif(PLATFORM STREQUAL "inference")
find_module(slog libslog.so ${ASCEND_ATC_DIR})
find_module(mmpa libmmpa.so ${ASCEND_ATC_DIR})
find_module(adump_server libadump_server.a ${ASCEND_ACL_DIR})
find_module(runtime libruntime.so ${ASCEND_ACL_DIR})
find_module(runtime_compile libruntime_compile.so ${ASCEND_ATC_DIR})
@@ -94,21 +93,21 @@ if (ENABLE_OPEN_SRC)
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR})
find_module(msprof libmsprof.so ${ASCEND_DRIVER_DIR})
else()
find_module(msprof libmsprof.so ${ASCEND_ATC_DIR})
find_module(msprof libmsprof.so ${ASCEND_DRIVER_SHARE_DIR})
endif()
elseif(PLATFORM STREQUAL "all")
find_module(slog libslog.so ${ASCEND_DRIVER_DIR})
find_module(mmpa libmmpa.so ${ASCEND_DRIVER_DIR})
find_module(msprof libmsprof.so ${ASCEND_DRIVER_DIR})
#mdc没有 dc在driver目录
find_module(msprof libmsprof.so ${ASCEND_DRIVER_COMMON_DIR})
find_module(hccl libhccl.so ${ASCEND_RUNTIME_DIR})
find_module(adump_server libadump_server.a ${ASCEND_RUNTIME_DIR})
find_module(runtime libruntime.so ${ASCEND_RUNTIME_DIR})
find_module(adump_server libadump_server.a ${ASCEND_ACL_DIR})
find_module(runtime libruntime.so ${ASCEND_ACL_DIR})
find_module(runtime_compile libruntime_compile.so ${ASCEND_ATC_DIR})
find_module(resource libresource.so ${ASCEND_RUNTIME_DIR})
find_module(error_manager liberror_manager.so ${ASCEND_RUNTIME_DIR})
find_module(resource libresource.so ${ASCEND_ATC_DIR})
find_module(error_manager liberror_manager.so ${ASCEND_ATC_DIR})
find_module(error_manager_static liberror_manager.a ${ASCEND_ACL_DIR})
find_module(msprofiler libmsprofiler.a ${ASCEND_RUNTIME_DIR})
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR})
find_module(msprofiler libmsprofiler.a ${ASCEND_ACL_DIR})
#mdc没有 dc在driver/lib64/driver
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}/driver)
find_module(ascendcl_static libascendcl.a ${ASCEND_ACL_DIR})
else()
message(FATAL_ERROR "PLATFORM param is invalid, should be train or inference, build terminated")


+ 2
- 2
cmake/external_libs/json.cmake View File

@@ -4,11 +4,11 @@ endif()

include(ExternalProject)

set(JSON_SRC_DIR ${GE_CODE_DIR}/../third_party/json/include)
set(JSON_SRC_DIR ${CMAKE_BINARY_DIR}/opensrc/json/include)
ExternalProject_Add(json_build
URL https://github.com/nlohmann/json/releases/download/v3.6.1/include.zip
#URL /home/txd/workspace/cloud_code/pkg/include.zip
#SOURCE_DIR ${JSON_SRC_DIR}
SOURCE_DIR ${JSON_SRC_DIR}
CONFIGURE_COMMAND ""
BUILD_COMMAND ""
INSTALL_COMMAND ""


+ 27
- 0
ge/ge_runtime/proto/task.pb.h View File

@@ -0,0 +1,27 @@
/**
* Copyright 2020 Huawei Technologies Co., Ltd
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/

// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: task.proto

#ifndef STUB_TASK_PROTO_H
#define STUB_TASK_PROTO_H

namespace domi {
class TaskDef;
}

#endif // STUB_TASK_PROTO_H

+ 29
- 3
ge/generator/ge_generator.cc View File

@@ -528,9 +528,19 @@ Status GeGenerator::GenerateModel(const Graph &graph, const string &file_name_pr
return SUCCESS;
}

Status GeGenerator::BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &inputs, const vector<GeTensor> &outputs,
const string &model_file_name, OpEngineType engine_type, ModelBufferData &model_buff,
bool is_offline) {
namespace {
bool IsNeedConnectInputOpForSingleOp(GeTensorDesc &tensor_desc) {
bool is_need = true;
// format and dtype is all reserved, stand for Optional input. When singleop scene
if (tensor_desc.GetFormat() == FORMAT_RESERVED && tensor_desc.GetDataType() == DT_UNDEFINED) {
is_need = false;
}
return is_need;
}
}

Status GeGenerator::CheckForSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &inputs,
const vector<GeTensor> &outputs) {
GE_CHECK_NOTNULL_EXEC(op_desc, return PARAM_INVALID);
if (!inputs.empty() && (inputs.size() != op_desc->GetAllInputsSize())) {
GELOGE(PARAM_INVALID, "Tensor size: %zu, Inputs size: %zu", inputs.size(), op_desc->GetAllInputsSize());
@@ -540,7 +550,17 @@ Status GeGenerator::BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &in
GELOGE(PARAM_INVALID, "Tensor size: %zu, Outputs size: %zu", outputs.size(), op_desc->GetOutputsSize());
return PARAM_INVALID;
}
return SUCCESS;
}

Status GeGenerator::BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &inputs, const vector<GeTensor> &outputs,
const string &model_file_name, OpEngineType engine_type, ModelBufferData &model_buff,
bool is_offline) {

if (CheckForSingleOp(op_desc, inputs, outputs) != SUCCESS) {
GELOGE(PARAM_INVALID, "input param is invalid when build single op!");
return PARAM_INVALID;
}
OmgContext &omg_context = (impl_ == nullptr) ? domi::GetContext() : impl_->omg_context_;
omg_context.is_dynamic_input = ContainsDynamicInpus(*op_desc);

@@ -575,12 +595,18 @@ Status GeGenerator::BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &in
if (inputs.empty()) {
for (const auto &input_desc : op_desc->GetAllInputsDescPtr()) {
GE_CHECK_NOTNULL_EXEC(input_desc, return INTERNAL_ERROR);
if (!IsNeedConnectInputOpForSingleOp(*input_desc)) {
continue;
}
GE_CHK_STATUS_RET_NOLOG(AddInputs(compute_graph, op_node, *input_desc, arg_index, false));
arg_index++;
}
} else {
for (const auto &in_desc : inputs) {
GeTensorDesc input_desc = in_desc.GetTensorDesc();
if (!IsNeedConnectInputOpForSingleOp(input_desc)) {
continue;
}
GE_CHK_STATUS_RET_NOLOG(AddInputs(compute_graph, op_node, input_desc, arg_index, true));
arg_index++;
}


+ 14
- 2
ge/graph/build/memory/block_mem_assigner.cc View File

@@ -545,7 +545,8 @@ bool CanReuseBySize(const map<string, uint64_t> &reusable_block_counts, const Me
}

bool BlockMemAssigner::IsOutNodeSetContinuousInput(const NodePtr &n, uint32_t out_index, std::string &peer_name,
uint32_t &peer_input_index, bool &no_need_assign_memory) {
uint32_t &peer_input_index,
bool &no_need_assign_memory, bool &reset_zero_copy_flag) {
if (n == nullptr || n->GetAllOutDataAnchors().size() <= 0) {
return false;
}
@@ -571,6 +572,13 @@ bool BlockMemAssigner::IsOutNodeSetContinuousInput(const NodePtr &n, uint32_t ou
return false;);

// If GetBool fail, is_input_continuous is false.
bool is_input_continuous_no_padding = false;
(void)ge::AttrUtils::GetBool(peer_in_node_desc, ATTR_NAME_NOPADDING_CONTINUOUS_INPUT,
is_input_continuous_no_padding);
if (is_input_continuous_no_padding) {
reset_zero_copy_flag = true;
return false;
}
(void)ge::AttrUtils::GetBool(peer_in_node_desc, ATTR_NAME_CONTINUOUS_INPUT, is_input_continuous);

GE_IF_BOOL_EXEC(is_input_continuous && CheckIsZeroMemNodeType(peer_node->GetType()),
@@ -1249,10 +1257,11 @@ Status BlockMemAssigner::AssignOutputMemoryWithReuse(const NodePtr &node, vector
std::string peer_name;
uint32_t peer_input_index = 0;
bool out_node_set_continuous_input = false;
bool reset_zero_copy_flag = false;
bool no_need_assign_memory = ((size == 0) || CheckIsZeroMemNodeType(node->GetType()));
if (!no_need_assign_memory) {
out_node_set_continuous_input =
IsOutNodeSetContinuousInput(node, i, peer_name, peer_input_index, no_need_assign_memory);
IsOutNodeSetContinuousInput(node, i, peer_name, peer_input_index, no_need_assign_memory, reset_zero_copy_flag);
GE_IF_BOOL_EXEC(!no_need_assign_memory,
no_need_assign_memory = IsAtomicOutputMemory(node, i, is_atomic, out_node_set_continuous_input););
}
@@ -1269,6 +1278,9 @@ Status BlockMemAssigner::AssignOutputMemoryWithReuse(const NodePtr &node, vector

MemoryBlock *mem_block = ApplyOutMemory(node, i, ranges, is_op_reuse_mem_, out_node_set_continuous_input);
if (mem_block != nullptr) {
GE_IF_BOOL_EXEC(reset_zero_copy_flag,
mem_block->is_zero_copy_ = false;
GELOGI("Node[%s] output[%u] need assign memory before reassign.", op_desc->GetName().c_str(), i););
node_out_blocks_[node->GetName()].emplace_back(mem_block);
if (out_node_set_continuous_input) {
node_continuous_input_blocks_[peer_name][peer_input_index] = mem_block;


+ 1
- 1
ge/graph/build/memory/block_mem_assigner.h View File

@@ -390,7 +390,7 @@ class BlockMemAssigner : public MemAssigner {
bool IsZeroCopyBlock(const NodePtr &node, bool continuous);

bool IsOutNodeSetContinuousInput(const NodePtr &n, uint32_t out_index, std::string &peer_name,
uint32_t &peer_input_index, bool &no_need_assign_memory);
uint32_t &peer_input_index, bool &no_need_assign_memory, bool &reset_zero_copy_flag);

///
/// @ingroup GE


+ 11
- 6
ge/ir_build/ge_ir_build.cc View File

@@ -171,7 +171,7 @@ class Impl {
graphStatus InitDomiOmgContext(const string &input_shape, const string &input_format, const string &net_format,
bool is_dynamic_input);
void SetRtSocVersion();
void UpdateThreadContext();
public:
ge::GeGenerator generator_;
std::map<std::string, std::string> options_;
@@ -225,8 +225,6 @@ graphStatus Impl::Init(const std::map<std::string, std::string> &options) {
return ret;
}

GetThreadLocalContext().SetGlobalOption(GetMutableGlobalOptions());
GetThreadLocalContext().SetGraphOption(options_);
std::string build_mode = (options_.find(BUILD_MODE) == options_.end() || options_[BUILD_MODE] == BUILD_MODE_NORMAL)
? "" : options_[BUILD_MODE];
options_[BUILD_MODE] = build_mode;
@@ -286,7 +284,7 @@ graphStatus Impl::Init(const std::map<std::string, std::string> &options) {
ge::PrintOptionMap(options_, "ge option");

SetRtSocVersion();
UpdateThreadContext();
// 3. init generator with options_
ret = generator_.Initialize(options_, omg_context_);
if (ret != GRAPH_SUCCESS) {
@@ -310,6 +308,11 @@ void Impl::SetRtSocVersion() {
}
}

void Impl::UpdateThreadContext() {
GetThreadLocalContext().SetGlobalOption(GetMutableGlobalOptions());
GetThreadLocalContext().SetGraphOption(options_);
}

graphStatus Impl::CreateInputsForIRBuild(const ge::Graph &graph, vector<ge::GeTensor> &inputs) {
auto compute_graph = ge::GraphUtils::GetComputeGraph(graph);
GE_CHECK_NOTNULL(compute_graph);
@@ -333,13 +336,15 @@ graphStatus Impl::CreateInputsForIRBuild(const ge::Graph &graph, vector<ge::GeTe
data_shape = tensor.GetShape();
GELOGI("Data op get shape from InputDesc in ge ir graph.");
}

// If user point input format, do work for all data ops; else do according to tensor_desc
auto data_format = omg_context_.format != domi::DOMI_TENSOR_ND ?
ge::TypeUtils::DomiFormatToFormat(omg_context_.format) : tensor.GetFormat();
ge::DataType data_type = tensor.GetDataType();
string data_type_str = ge::TypeUtils::DataTypeToSerialString(data_type);
GELOGI("Data op get data type:%s from InputDesc in ge ir graph.", data_type_str.c_str());

ge::GeTensor inputTensor;
ge::GeTensorDesc desc(data_shape, ge::Format(omg_context_.format), data_type);
ge::GeTensorDesc desc(data_shape, ge::Format(data_format), data_type);
inputTensor.SetTensorDesc(desc);
inputs.push_back(inputTensor);
}


+ 5
- 10
ge/offline/single_op_parser.cc View File

@@ -226,16 +226,11 @@ bool SingleOpParser::Validate(const SingleOpDesc &op_desc) {
}

int index = 0;
for (auto &tensor_desc : op_desc.input_desc) {
if (tensor_desc.type == DT_UNDEFINED) {
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "index"}, {"input", std::to_string(index)});
GELOGE(false, "Input's dataType is invalid when the index is %d", index);
return false;
}

if (tensor_desc.format == FORMAT_RESERVED) {
ErrorManager::GetInstance().ATCReportErrMessage("E10028", {"input", "index"}, {"input", std::to_string(index)});
GELOGE(PARAM_INVALID, "Input's format is invalid when the index is %d", index);
for (auto &tensor_desc : op_desc.output_desc) {
if ((tensor_desc.type == DT_UNDEFINED && tensor_desc.format != FORMAT_RESERVED) ||
(tensor_desc.type != DT_UNDEFINED && tensor_desc.format == FORMAT_RESERVED)){
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "index"}, {"output", std::to_string(index)});
GELOGE(PARAM_INVALID, "Input's dataType or format is invalid when the index is %d", index);
return false;
}
++index;


+ 8
- 8
inc/framework/generator/ge_generator.h View File

@@ -53,7 +53,7 @@ class GeGenerator {
Status GenerateOfflineModel(const Graph &graph, const std::string &file_name_prefix,
const std::vector<GeTensor> &inputs = std::vector<GeTensor>());

Status GenerateOnlineModel(const Graph &graph, const vector<GeTensor> &inputs, ge::ModelBufferData& model);
Status GenerateOnlineModel(const Graph &graph, const vector<GeTensor> &inputs, ge::ModelBufferData &model);

Status GenerateInfershapeGraph(const Graph &graph);

@@ -77,16 +77,16 @@ class GeGenerator {
/// @param [in] engine_type: specific engine.
/// @param [out] model_buff: model buff of single op.
/// @return SUCCESS or FAILED
Status BuildSingleOpModel(OpDescPtr &op_desc, const vector<GeTensor> &inputs,
const vector<GeTensor> &outputs, OpEngineType engine_type,
ModelBufferData &model_buff);
Status BuildSingleOpModel(OpDescPtr &op_desc, const vector<GeTensor> &inputs, const vector<GeTensor> &outputs,
OpEngineType engine_type, ModelBufferData &model_buff);

private:
Status GenerateModel(const Graph &graph, const string &file_name_prefix,
const vector<GeTensor> &inputs, ge::ModelBufferData& model, bool is_offline = true);
Status GenerateModel(const Graph &graph, const string &file_name_prefix, const vector<GeTensor> &inputs,
ge::ModelBufferData &model, bool is_offline = true);
Status BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &inputs, const vector<GeTensor> &outputs,
const string &model_file_name, OpEngineType engine_type,
ModelBufferData &model_buff, bool is_offline = true);
const string &model_file_name, OpEngineType engine_type, ModelBufferData &model_buff,
bool is_offline = true);
Status CheckForSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &inputs, const vector<GeTensor> &outputs);

class Impl;



+ 1
- 1
metadef

@@ -1 +1 @@
Subproject commit e79f45025b95a7295bdffa38a1b720bfd6609ae9
Subproject commit 1cc55bcae09902b3d158993dd57bfbd1d3337066

+ 1
- 1
parser

@@ -1 +1 @@
Subproject commit be653a17038395f3901d43b600a51f6ab33af5e5
Subproject commit db4e6070bb2cec01cead264a44ceae07e7f3048e

+ 0
- 1
tests/depends/cce/src/cce_stub.cc View File

@@ -528,7 +528,6 @@ uint32_t Fusion(ComputeGraphPtr model_graph, ComputeGraphPtr fusion_graph, kScop
int stream_num = 1;
int flag = 0;

// make_graph_nd(graph);
NodePtr node_a = fusion_graph->AddNode(op_def_a);
NodePtr node_b = fusion_graph->AddNode(op_def_b);



+ 0
- 2
tests/st/resnet50/resnet50_train.cc View File

@@ -746,7 +746,6 @@ int TestBuildGraphTest(Func fun, Graph &graph, vector<ge::Tensor> &inputs, vecto
shapeTensor.SetTensorDesc(shape_desc);
vector<float> dataValuec;
for (int i = 0; i < sizeshape; i++) {
// dataValuec.push_back((float)(i%255));
dataValuec.push_back(1);
}

@@ -764,7 +763,6 @@ int TestBuildGraphTest(Func fun, Graph &graph, vector<ge::Tensor> &inputs, vecto
}

shapeTensor1.SetData((uint8_t *)dataValuec1.data(), 4 * sizeshape1);
// inputs.push_back(shapeTensor1);

return 0;
}


+ 0
- 2
tests/ut/common/graph/testcase/ge_graph/ge_model_unittest.cc View File

@@ -69,12 +69,10 @@ TEST_F(UtestGeModelUnittest, save_model_to_file_success) {
ge::Graph ge_graph = ge::GraphUtils::CreateGraphFromComputeGraph(compute_graph);
string file_name = "model_data.pb";
setenv("DUMP_MODEL", "1", true);
// EXPECT_EQ(ge_graph.SaveToFile(file_name), GRAPH_FAILED);
setenv("DUMP_MODEL", "0", true);
}

TEST_F(UtestGeModelUnittest, load_model_from_file_success) {
ge::Graph ge_graph;
string file_name = "model_data.pb";
// EXPECT_EQ(ge_graph.LoadFromFile(file_name), GRAPH_SUCCESS);
}

+ 0
- 16
tests/ut/ge/graph/load/new_model_manager_davinci_model_unittest.cc View File

@@ -182,8 +182,6 @@ TEST_F(UtestModelManagerDavinciModel, contruct_modeldef_createfail) {
ge::AttrUtils::SetListInt(op_desc, ge::POOLING_ATTR_WINDOW, vector<int>({1, 1}));
ge::AttrUtils::SetListInt(op_desc, ge::POOLING_ATTR_STRIDE, vector<int>({1, 1}));

// EXPECT_EQ(ge::SUCCESS, model.Init());

model.GetEventList();
}

@@ -200,7 +198,6 @@ TEST_F(UtestModelManagerDavinciModel, copy_input_data_to_model_fail) {
input_data.blobs.push_back(data_buffer);

model.op_list_.clear();
// EXPECT_EQ(ge::PARAM_INVALID, model.CopyInputDataToModel(input_data.blobs, 0));

delete[](char *) data_buffer.data;
}
@@ -210,7 +207,6 @@ TEST_F(UtestModelManagerDavinciModel, streamnum_success) {
DavinciModel *model = new DavinciModel(0, g_label_call_back);

OmeTestOpUtils::InitModel(*model);
// EXPECT_EQ(ge::SUCCESS, model->Init());

EXPECT_EQ(0, model->StreamNum());
EXPECT_EQ(ge::INTERNAL_ERROR, model->ModelRunStart());
@@ -226,8 +222,6 @@ TEST_F(UtestModelManagerDavinciModel, eventnum_success) {

OmeTestOpUtils::InitModel(*model);

// EXPECT_EQ(ge::SUCCESS, model->Init());

EXPECT_EQ(0, model->EventNum());
EXPECT_EQ(ge::INTERNAL_ERROR, model->ModelRunStart());

@@ -241,8 +235,6 @@ TEST_F(UtestModelManagerDavinciModel, handlelist_success) {

OmeTestOpUtils::InitModel(*model);

// EXPECT_EQ(ge::SUCCESS, model->Init());

EXPECT_EQ(ge::INTERNAL_ERROR, model->ModelRunStart());

EXPECT_EQ(ge::SUCCESS, model->ModelRunStop());
@@ -256,8 +248,6 @@ TEST_F(UtestModelManagerDavinciModel, eventlist_success) {

OmeTestOpUtils::InitModel(*model);

// EXPECT_EQ(ge::SUCCESS, model->Init());

EXPECT_EQ(true, model->GetEventList().empty());
EXPECT_EQ(ge::INTERNAL_ERROR, model->ModelRunStart());

@@ -282,7 +272,6 @@ TEST_F(UtestModelManagerDavinciModel, failed_reset_device) {
TEST_F(UtestModelManagerDavinciModel, init_not_support_priority) {
int32_t priority = 8;
DavinciModel model(priority, g_label_call_back);
// EXPECT_EQ(ge::PARAM_INVALID, model.Init());
}

// test GetInputOutputDescInfo
@@ -346,7 +335,6 @@ TEST_F(UtestModelManagerDavinciModel, CopyTensorFromSrcVarNode_success) {
NodePtr dst_node = graph->AddNode(op_desc_ptr);
DavinciModel model(0, g_label_call_back);
Status ret = model.CopyTensorFromSrcVarNode(src_node, dst_node);
// EXPECT_EQ(SUCCESS, ret);
}

TEST_F(UtestModelManagerDavinciModel, CopyVarData_graph_is_nullptr) {
@@ -370,7 +358,6 @@ TEST_F(UtestModelManagerDavinciModel, copy_var_data_success) {

DavinciModel model(0, g_label_call_back);
Status ret = model.CopyVarData(graph);
// EXPECT_EQ(SUCCESS, ret);
}

TEST_F(UtestModelManagerDavinciModel, get_input_output_desc_info_without_data_op_list) {
@@ -540,7 +527,6 @@ TEST_F(UtestModelManagerDavinciModel, get_flow_ctrl_op_list_success) {
std::map<uint32_t, uint32_t> flowctrl_op_index_internal_map;
flowctrl_op_index_internal_map.insert(pair<uint32_t, uint32_t>(1, 1));
model.flowctrl_op_index_internal_map_ = flowctrl_op_index_internal_map;
// EXPECT_EQ(flowctrl_op_index_internal_map_, model.GetFlowctrlOpList());
}

// test SetFlowctrlOpList
@@ -1204,10 +1190,8 @@ TEST_F(UtestModelManagerDavinciModel, profiling_model_success) {
input_data.index = 0;
input_data.model_id = 1;
input_data.blobs.push_back(data_buffer);
// model.SinkModelProfile(&model);

rtFreeHost(data.model_data);
// delete stream;
delete[](char *) data_buffer.data;
delete model_def;
}


+ 0
- 95
tests/ut/ge/graph/load/new_model_manager_model_manager_unittest.cc View File

@@ -153,20 +153,6 @@ TEST_F(UtestModelManagerModelManager, case_load_model_encypt_not_match) {
delete[](uint8_t *) data.model_data;
}

#if 0
TEST_F(UtestModelManagerModelManager, case_load_model_signature_failed)
{
ModelManager mm;
ge::ModelData data;
GenUnencryptModelData(data);

uint32_t model_id = 1;
MOCKER(&WBDecryptor::CheckSignature).stubs().will(returnValue(false));
EXPECT_EQ(ge::PARAM_INVALID, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}
#endif

TEST_F(UtestModelManagerModelManager, case_load_model_encypt_type_unsupported) {
ModelManager mm;
ge::ModelData data;
@@ -178,87 +164,6 @@ TEST_F(UtestModelManagerModelManager, case_load_model_encypt_type_unsupported) {
delete[](uint8_t *) data.model_data;
}

#if 0
TEST_F(UtestModelManagerModelManager, case_load_model_header_len_failed)
{
ModelManager mm;
ge::ModelData data;
GenEncryptModelData(data);
ModelFileHeader *header = (ModelFileHeader*)data.model_data;
data.model_len -= header->length;
header->length = 0;
uint32_t model_id = 1;
EXPECT_EQ(ge::PARAM_INVALID, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}
#endif

#if 0
TEST_F(UtestModelManagerModelManager, case_load_success)
{
const char* model_file = "bin/llt/framework/domi/ut/omg/data/leakyrelu.dav";
const char* json_file = "test.json";
const char* key = "bin/llt/framework/domi/ut/omg/data/leakyrelu.dav.PASSCODE";

ge::ModelData model;
Status ret = ModelParserBase::LoadFromFile(model_file, key, 0, &model);
EXPECT_EQ(ge::SUCCESS, ret);

ModelManager mm;
uint32_t model_id = 1;
ret = mm.LoadModelOffline(model_id, model, UTEST_CALL_BACK_FUN);
EXPECT_EQ(ge::SUCCESS, ret);

if (model.model_data)
delete[](uint8_t*)model.model_data;
}
#endif

#if 0
TEST_F(UtestModelManagerModelManager, case_load_encrypt_model_signature_failed)
{
ModelManager mm;
ge::ModelData data;
GenEncryptModelData(data);
uint32_t model_id = 1;
data.key;
EXPECT_EQ(ge::PARAM_INVALID, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}

TEST_F(UtestModelManagerModelManager, case_load_encrypt_model_invalid_key_len)
{
ModelManager mm;
ge::ModelData data;
GenEncryptModelData(data);
data.key = "0123456789abcdef0123456789ABCDEF0123456789ABCDEF0123456789ABCDEF0";
uint32_t model_id = 1;
EXPECT_EQ(ge::PARAM_INVALID, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}

TEST_F(UtestModelManagerModelManager, case_load_encrypt_model_invalid_key_char)
{
ModelManager mm;
ge::ModelData data;
GenEncryptModelData(data);
data.key = "0123456789abcdef0123456789ABCDEF0123456789ABCDEF0123456789ABCDEG";
uint32_t model_id = 1;
EXPECT_EQ(ge::PARAM_INVALID, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}

TEST_F(UtestModelManagerModelManager, case_load_encrypt_model_load_failed)
{
ModelManager mm;
ge::ModelData data;
GenEncryptModelData(data);
uint32_t model_id = 1;
EXPECT_EQ(ge::INTERNAL_ERROR, mm.LoadModelOffline(model_id, data, UTEST_CALL_BACK_FUN));
delete[](uint8_t*)data.model_data;
}
#endif

shared_ptr<ge::ModelListener> LabelCallBack(new DModelListener());

// test HandleCommand


+ 0
- 5
tests/ut/ge/graph/load/new_op_test_utils.h View File

@@ -76,7 +76,6 @@ class OmeTestOpUtils {
return nullptr;
}

// return std::make_shared<ge::Node>(op_desc, nullptr);
auto g = std::make_shared<ge::ComputeGraph>("g");
return g->AddNode(std::move(op_desc));
}
@@ -403,8 +402,6 @@ class OmeTestOpDescBuilder {
if (SUCCESS != res) {
GELOGE(ge::FAILED, "Finish: GraphUtils::AddEdge failed");
}
// ge::NodePtr src_node = node->GetOwnerComputeGraph()->AddNodeFront(src_op_desc);
// node->AddLinkFrom(src_node);
}

{
@@ -434,8 +431,6 @@ class OmeTestOpDescBuilder {
vector<ge::GeTensorPtr> weights_;
int64_t eventId_ = -1;
int64_t scopeid_ = -1;

// std::shared_ptr<ge::ComputeGraph> graph_;
};

#endif // OME_REBUILD_OME_OP_TEST_UTILS_H

+ 0
- 1
tests/ut/ge/graph/passes/dimension_adjust_pass_unittest.cc View File

@@ -122,7 +122,6 @@ TEST_F(UtestGraphPassesDimensionAdjustPass, node_get_original_type_failed) {

std::shared_ptr<DimensionAdjustPass> pass = make_shared<DimensionAdjustPass>();
ge::Status ret = pass->Run(op_node);
// EXPECT_EQ(ge::SUCCESS, ret);
}

TEST_F(UtestGraphPassesDimensionAdjustPass, node_not_register_op) {


+ 0
- 15
tests/ut/ge/graph/passes/folding_kernel/strided_slice_kernel_unittest.cc View File

@@ -93,7 +93,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test2) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test3) {
@@ -123,7 +122,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test3) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test4) {
@@ -154,7 +152,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test4) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test5) {
@@ -186,7 +183,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test5) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test6) {
@@ -219,7 +215,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test6) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test7) {
@@ -253,7 +248,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test7) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test8) {
@@ -288,7 +282,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test8) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test9) {
@@ -322,7 +315,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test9) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test10) {
@@ -357,7 +349,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test10) {
shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);

ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test11) {
@@ -392,7 +383,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test11) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test12) {
@@ -427,7 +417,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test12) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test13) {
@@ -462,7 +451,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test13) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test14) {
@@ -497,7 +485,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test14) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test15) {
@@ -532,7 +519,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test15) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test16) {
@@ -567,7 +553,6 @@ TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test16) {

shared_ptr<Kernel> kernel = KernelFactory::Instance().Create(STRIDEDSLICE);
ge::Status status = kernel->Compute(op_desc_ptr, input, outputs);
// EXPECT_EQ(PARAM_INVALID, status);
}

TEST_F(UtestGraphPassesFoldingKernelStridedSliceKernel, Test17) {


+ 0
- 1
tests/ut/ge/graph/passes/guarantee_const_pass_unittest.cc View File

@@ -167,7 +167,6 @@ TEST_F(UtestGraphPassesGuaranteeConstPass, get_origenal_type_fail) {
string type2 = "FrameworkOp";
node->GetOpDesc()->SetType(type2);
ge::Status ret = guarantee_const_op_remove_pass_->Run(node);
// EXPECT_EQ(ge::SUCCESS, ret);
}

TEST_F(UtestGraphPassesGuaranteeConstPass, int32_success_6) {


+ 0
- 1
tests/ut/ge/graph/passes/identity_pass_unittest.cc View File

@@ -135,7 +135,6 @@ TEST_F(UtestIdentityPass, succ) {
string type2 = "FrameworkOp";
node->GetOpDesc()->SetType(type2);
status = pass.Run(node);
// EXPECT_EQ(ge::SUCCESS, status);

NodePtr node_err = AddNode(graph, "Identity", IDENTITY, 1, 2);
status = pass.Run(node_err);


+ 0
- 1
tests/ut/ge/graph/passes/net_output_pass_unittest.cc View File

@@ -845,7 +845,6 @@ TEST_F(UtestGraphPassesNetOutputPass, out_node_remove_check_fail) {
ge::NodePtr mul2 = compute_graph->FindNode("Mul2");
std::vector<std::pair<ge::NodePtr, int32_t>> output_nodes = {{mul1, 0}, {mul2, 0}};
compute_graph->SetGraphOutNodesInfo(output_nodes);
// compute_graph->RemoveNode(mul1);
mul1->GetInDataAnchor(0)->UnlinkAll();
mul1->GetInDataAnchor(1)->UnlinkAll();
GraphUtils::RemoveNodeWithoutRelink(compute_graph, mul1);


+ 0
- 1
tests/ut/ge/graph/passes/placeholder_with_default_pass_unittest.cc View File

@@ -75,5 +75,4 @@ TEST_F(UtestPlaceholderWithDefaultPass, succ) {
string type2 = "FrameworkOp";
node->GetOpDesc()->SetType(type2);
pass.Run(node);
// EXPECT_EQ(ge::SUCCESS, status);
}

+ 0
- 1
tests/ut/ge/graph/passes/prevent_gradient_pass_unittest.cc View File

@@ -75,5 +75,4 @@ TEST_F(UtestPreventGradientPass, succ) {
string type2 = "FrameworkOp";
node->GetOpDesc()->SetType(type2);
status = pass.Run(node);
// EXPECT_EQ(ge::SUCCESS, status);
}

+ 0
- 1
tests/ut/ge/graph/passes/reshape_remove_pass_unittest.cc View File

@@ -178,6 +178,5 @@ TEST_F(UtestReshapeRemovePass, reshape_remove_without_const) {
EXPECT_EQ(var1->GetOutDataNodes().at(0)->GetName(), "transdata1");
EXPECT_NE(const1, nullptr);
EXPECT_EQ(const1->GetOutNodes().size(), 1);
// EXPECT_EQ(const1->GetOutDataNodes().at(0)->GetName(), "transdata2");
}
} // namespace ge

+ 0
- 1
tests/ut/ge/graph/passes/snapshot_pass_unittest.cc View File

@@ -78,5 +78,4 @@ TEST_F(UtestSnapshotPass, succ) {
string type2 = "FrameworkOp";
snapshot->GetOpDesc()->SetType(type2);
status = pass.Run(snapshot);
// EXPECT_EQ(ge::SUCCESS, status);
}

+ 0
- 1
tests/ut/ge/graph/passes/stop_gradient_pass_unittest.cc View File

@@ -176,7 +176,6 @@ TEST_F(UtestGraphPassesStopGradientPass, get_origenal_type_fail) {
string type2 = "FrameworkOp";
node->GetOpDesc()->SetType(type2);
ge::Status ret = pass_->Run(node);
// EXPECT_EQ(ge::SUCCESS, ret);
}
TEST_F(UtestGraphPassesStopGradientPass, size_check_fail) {
vector<int64_t> dims_vec_0 = {8, 2};


+ 0
- 2
tests/ut/ge/graph/passes/switch_pass_unittest.cc View File

@@ -256,7 +256,6 @@ TEST_F(UtestGraphPassesSwitchPass, inactive_output_not_exists) {
output_true_node_->GetOutDataAnchor(0)->UnlinkAll();
GraphUtils::RemoveNodeWithoutRelink(graph_, output_true_node_);
switch_node_->GetOutDataAnchor(1)->UnlinkAll();
// switch_node_->outDataAnchors_.pop_back();

/// input
/// |
@@ -394,7 +393,6 @@ TEST_F(UtestGraphPassesSwitchPass, dead_output_connected_to_merge) {
/// Merge
bool pred_value = true;
BuildDefaultGraph(false, &pred_value);
// graph_->RemoveNode(output_false_node_);
output_false_node_->GetOutDataAnchor(0)->UnlinkAll();
GraphUtils::RemoveNodeWithoutRelink(graph_, output_false_node_);
switch_node_->GetOutDataAnchor(0)->UnlinkAll();


+ 0
- 1
tests/ut/ge/graph/passes/unused_and_isolated_op_remove_pass_unittest.cc View File

@@ -106,7 +106,6 @@ TEST_F(UtestGraphPassesUnusedAndIsolatedOpRemovePass, transpose_and_conv) {
Status status = PassManager::Run(graph, passes);
EXPECT_EQ(SUCCESS, status);
NodePtr found_node0 = graph->FindNode("transpose1");
// EXPECT_EQ(nullptr, found_node0);
NodePtr found_node = graph->FindNode("conv1");
EXPECT_EQ(conv_node, found_node);
}


+ 0
- 4
tests/ut/ge/graph/passes/variable_op_pass_unittest.cc View File

@@ -343,8 +343,6 @@ bool BuildComputeGraph0(ge::ComputeGraphPtr &graph) {

if (ge::GraphUtils::AddEdge(node_apply_monetum->GetOutDataAnchor(0), node_5d_to_4d_1->GetInDataAnchor(0)) !=
ge::SUCCESS) {
/// GELOGE(FAILED, "ge::GraphUtils::AddEdge(node_apply_monetum->GetOutDataAnchor(0),
/// node_5d_to_4d_1->GetInDataAnchor(0) ) Failed.");
};
ge::GraphUtils::AddEdge(node_5d_to_4d_1->GetOutDataAnchor(0), node_ref->GetInDataAnchor(0));

@@ -395,8 +393,6 @@ bool BuildComputeGraph1(ge::ComputeGraphPtr &graph) {

if (ge::GraphUtils::AddEdge(node_apply_monetum->GetOutDataAnchor(0), node_5d_to_4d_1->GetInDataAnchor(0)) !=
ge::SUCCESS) {
/// GELOGE(FAILED, "ge::GraphUtils::AddEdge(node_apply_monetum->GetOutDataAnchor(0),
/// node_5d_to_4d_1->GetInDataAnchor(0) ) Failed.");
};
ge::GraphUtils::AddEdge(node_5d_to_4d_1->GetOutDataAnchor(0), node_ref->GetInDataAnchor(0));



Loading…
Cancel
Save