You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

graph_builder.cc 22 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492
  1. /**
  2. * Copyright 2019-2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "graph/build/graph_builder.h"
  17. #include "common/ge/ge_util.h"
  18. #include "common/helper/model_helper.h"
  19. #include "common/opskernel/ops_kernel_info_types.h"
  20. #include "graph/build/run_context.h"
  21. #include "graph/build/stream_graph_optimizer.h"
  22. #include "graph/manager/graph_var_manager.h"
  23. #include "graph/utils/node_utils.h"
  24. #include "graph/utils/type_utils.h"
  25. #include "init/gelib.h"
  26. #include "model/ge_model.h"
  27. using domi::BuildMode;
  28. namespace {
  29. const int32_t kInvalidPerfLevel = -1;
  30. } // namespace
  31. namespace ge {
  32. GraphBuilder::GraphBuilder() : build_mode_(BuildMode::GEN_TASK_WITH_FUSION), hcom_parallel_(false) {}
  33. void GraphBuilder::SetOptions(const ge::GraphManagerOptions &options) {
  34. stream_max_parallel_num_ = options.stream_max_parallel_num;
  35. hcom_parallel_ = options.hcom_parallel;
  36. if (options.perf_level == kInvalidPerfLevel) {
  37. build_mode_ = static_cast<int>(BuildMode::GEN_TASK_WITH_FUSION);
  38. } else {
  39. build_mode_ = options.perf_level;
  40. }
  41. }
  42. Status GraphBuilder::CalcOpParam(const ge::ComputeGraphPtr &graph) {
  43. GELOGI("Begin to calculate op running param.");
  44. GE_CHECK_NOTNULL(graph);
  45. auto instance_ptr = ge::GELib::GetInstance();
  46. if (instance_ptr == nullptr || !instance_ptr->InitFlag()) {
  47. GELOGE(GE_CLI_GE_NOT_INITIALIZED, "GraphBuilder: GE is not initialized");
  48. return GE_CLI_GE_NOT_INITIALIZED;
  49. }
  50. for (const auto &node_ptr : graph->GetAllNodes()) {
  51. GE_CHECK_NOTNULL(node_ptr->GetOpDesc());
  52. std::string kernel_lib_name = node_ptr->GetOpDesc()->GetOpKernelLibName();
  53. if (kernel_lib_name.empty()) {
  54. // reset op kernel lib
  55. (void)instance_ptr->DNNEngineManagerObj().GetDNNEngineName(node_ptr->GetOpDesc());
  56. kernel_lib_name = node_ptr->GetOpDesc()->GetOpKernelLibName();
  57. if (kernel_lib_name.empty()) {
  58. GELOGE(INTERNAL_ERROR, "Get node:%s(%s) kernel lib failed.", node_ptr->GetName().c_str(),
  59. node_ptr->GetType().c_str());
  60. return INTERNAL_ERROR;
  61. }
  62. }
  63. OpsKernelInfoStorePtr kernel_info = instance_ptr->OpsKernelManagerObj().GetOpsKernelInfoStore(kernel_lib_name);
  64. if (kernel_info != nullptr) {
  65. auto ret = SetInputSize(node_ptr);
  66. if (ret != SUCCESS) {
  67. GELOGE(ret, "Set node inputDesc size failed, node name is %s", node_ptr->GetName().c_str());
  68. return ret;
  69. }
  70. ret = kernel_info->CalcOpRunningParam(*node_ptr);
  71. if (ret != SUCCESS) {
  72. GELOGE(ret, "Calculate op running param failed, node name is %s", node_ptr->GetName().c_str());
  73. return ret;
  74. }
  75. } else {
  76. GELOGE(GE_GRAPH_PARAM_NULLPTR, "Get op %s ops kernel info store failed", node_ptr->GetName().c_str());
  77. return INTERNAL_ERROR;
  78. }
  79. }
  80. auto parent_node = graph->GetParentNode();
  81. if (parent_node == nullptr) {
  82. GELOGI("Graph[%s] do not have parent node, no need update parent node output size.", graph->GetName().c_str());
  83. return SUCCESS;
  84. }
  85. GE_CHK_STATUS_RET(UpdateParentNodeOutputSize(graph, parent_node));
  86. GELOGI("Success to calculate op running param.");
  87. return SUCCESS;
  88. }
  89. Status GraphBuilder::UpdateParentNodeOutputSize(const ge::ComputeGraphPtr &graph, ge::NodePtr &parent_node_ptr) {
  90. GELOGI("Begin to update parent node[%s] of graph[%s] output size.", parent_node_ptr->GetName().c_str(),
  91. graph->GetName().c_str());
  92. auto parent_op_desc = parent_node_ptr->GetOpDesc();
  93. GE_CHECK_NOTNULL(parent_op_desc);
  94. bool is_unknown_shape = false;
  95. if (!AttrUtils::GetBool(parent_op_desc, ATTR_NAME_IS_UNKNOWN_SHAPE, is_unknown_shape)) {
  96. GELOGE(PARAM_INVALID, "Get op %s unknown shape attr failed.", parent_op_desc->GetName().c_str());
  97. return PARAM_INVALID;
  98. }
  99. if (is_unknown_shape) {
  100. GELOGI("Current graph[%s] is unknown, no need to update parent node[%s] output size.", graph->GetName().c_str(),
  101. parent_node_ptr->GetName().c_str());
  102. return SUCCESS;
  103. }
  104. for (const auto &node_ptr : graph->GetDirectNode()) {
  105. if (node_ptr->GetType() != NETOUTPUT) {
  106. continue;
  107. }
  108. auto op_desc = node_ptr->GetOpDesc();
  109. GE_CHECK_NOTNULL(op_desc);
  110. for (const auto &in_data_anchor : node_ptr->GetAllInDataAnchors()) {
  111. auto index = in_data_anchor->GetIdx();
  112. ge::GeTensorDesc desc_temp = op_desc->GetInputDesc(index);
  113. int64_t size = 0;
  114. GE_IF_BOOL_EXEC(ge::TensorUtils::GetSize(desc_temp, size) != SUCCESS, GELOGI("Get size failed!"));
  115. uint32_t parent_index = 0;
  116. if (!AttrUtils::GetInt(desc_temp, ATTR_NAME_PARENT_NODE_INDEX, parent_index)) {
  117. GELOGE(INTERNAL_ERROR, "NetOutput input tensor %d, attr %s not found.", index,
  118. ATTR_NAME_PARENT_NODE_INDEX.c_str());
  119. return INTERNAL_ERROR;
  120. }
  121. ge::GeTensorDesc parent_desc_temp = parent_op_desc->GetOutputDesc(parent_index);
  122. ge::TensorUtils::SetSize(parent_desc_temp, size);
  123. GE_CHK_STATUS_RET(parent_op_desc->UpdateOutputDesc(parent_index, parent_desc_temp));
  124. GELOGI("Update parent node[%s] output index[%u] to size[%ld].", parent_node_ptr->GetName().c_str(), parent_index,
  125. size);
  126. }
  127. }
  128. return SUCCESS;
  129. }
  130. Status GraphBuilder::Build(ComputeGraphPtr &comp_graph, std::vector<SubGraphInfoPtr> &subgraph_ptr_list,
  131. GeRootModelPtr &ge_root_model_ptr, uint64_t session_id) {
  132. GELOGI("Start to build model.");
  133. if (comp_graph == nullptr) {
  134. GELOGE(GE_GRAPH_PARAM_NULLPTR, "Graph build comp_graph is null.");
  135. return GE_GRAPH_PARAM_NULLPTR;
  136. }
  137. ge_root_model_ptr = MakeShared<ge::GeRootModel>(comp_graph);
  138. if (ge_root_model_ptr == nullptr) {
  139. return MEMALLOC_FAILED;
  140. }
  141. GeModelPtr ge_model_ptr = nullptr;
  142. bool is_dynamic_shape = false;
  143. // To be compatible with the old process, do not verify the return value temporarily.
  144. (void)AttrUtils::GetBool(comp_graph, ATTR_NAME_DYNAMIC_SHAPE_PARTITIONED, is_dynamic_shape);
  145. if (is_dynamic_shape) {
  146. GE_CHK_STATUS_RET(
  147. BuildForDynamicShapeGraph(comp_graph, subgraph_ptr_list, ge_root_model_ptr, ge_model_ptr, session_id),
  148. "Build for dynamic shape graph failed.");
  149. return SUCCESS;
  150. }
  151. GE_CHK_STATUS_RET(BuildForKnownShapeGraph(comp_graph, subgraph_ptr_list, ge_model_ptr, session_id),
  152. "Build for known shape graph failed.");
  153. ge_root_model_ptr->SetSubgraphInstanceNameToModel(comp_graph->GetName(), ge_model_ptr);
  154. return SUCCESS;
  155. }
  156. Status GraphBuilder::BuildForKnownShapeGraph(ComputeGraphPtr &comp_graph,
  157. std::vector<SubGraphInfoPtr> &subgraph_ptr_list, GeModelPtr &ge_model_ptr,
  158. uint64_t session_id) {
  159. GELOGI("Begin to build known shape graph[%s].", comp_graph->GetName().c_str());
  160. Status ret = SecondPartition(comp_graph, subgraph_ptr_list);
  161. GE_CHK_STATUS_RET(ret, "Graph[%s] second partition Failed.", comp_graph->GetName().c_str());
  162. auto subgraph_map = graph_partitioner_.GetSubGraphMap();
  163. GE_TIMESTAMP_START(BuildSubgraph);
  164. ge::ModelBuilder builder(comp_graph, subgraph_map, stream_max_parallel_num_, hcom_parallel_, build_mode_);
  165. GE_DUMP(comp_graph, "BeforePreBuildModel");
  166. GE_TIMESTAMP_START(PreBuildModel);
  167. GE_CHK_STATUS_RET(builder.PreBuildModel(), "Graph[%s] builder PreBuildModel() return fail.",
  168. comp_graph->GetName().c_str());
  169. GE_TIMESTAMP_END(PreBuildModel, "GraphBuilder::PreBuildModel");
  170. GE_DUMP(comp_graph, "AfterPreBuildModel");
  171. GE_TIMESTAMP_START(CalcOpParam);
  172. GE_CHK_STATUS_RET(CalcOpParam(comp_graph), "Graph[%s] builder CalcOpParam() return fail.",
  173. comp_graph->GetName().c_str());
  174. GE_TIMESTAMP_END(CalcOpParam, "GraphBuilder::CalcOpParam");
  175. GE_DUMP(comp_graph, "AfterCalcOpParam");
  176. ModelPtr model_ptr = MakeShared<ge::Model>();
  177. if (model_ptr == nullptr) {
  178. return MEMALLOC_FAILED;
  179. }
  180. GE_TIMESTAMP_START(BuildModelForGetTask);
  181. GE_CHK_STATUS_RET(builder.BuildModelForGetTask(*model_ptr), "Graph[%s] builder BuildModelForGetTask() return fail.",
  182. comp_graph->GetName().c_str());
  183. GE_TIMESTAMP_END(BuildModelForGetTask, "GraphBuilder::BuildModelForGetTask");
  184. GE_DUMP(comp_graph, "AfterBuildModel");
  185. GE_TIMESTAMP_START(GetTaskInfo);
  186. ret = GetTaskInfo(builder, model_ptr, comp_graph, subgraph_map, session_id);
  187. GE_TIMESTAMP_END(GetTaskInfo, "GraphBuilder::GetTaskInfo");
  188. GE_DUMP(comp_graph, "AfterGetTask");
  189. if (ret != SUCCESS) {
  190. GELOGE(ret, "Graph[%s] builder GetTaskInfo() return fail.", comp_graph->GetName().c_str());
  191. return ret;
  192. }
  193. ge_model_ptr = MakeShared<ge::GeModel>();
  194. if (ge_model_ptr == nullptr) {
  195. return MEMALLOC_FAILED;
  196. }
  197. GE_CHK_STATUS_RET(builder.SaveDataToModel(*model_ptr, *ge_model_ptr),
  198. "Graph[%s] builder SaveDataToModel() return fail.", comp_graph->GetName().c_str());
  199. GELOGI("Success to build graph[%s] model.", comp_graph->GetName().c_str());
  200. GE_TIMESTAMP_END(BuildSubgraph, "GraphBuilder::Build");
  201. return SUCCESS;
  202. }
  203. Status GraphBuilder::BuildForUnknownShapeGraph(ComputeGraphPtr &comp_graph, GeModelPtr &ge_model_ptr,
  204. uint64_t session_id) {
  205. GELOGI("Begin to build unknown shape graph[%s].", comp_graph->GetName().c_str());
  206. GE_TIMESTAMP_START(CalcOpParam);
  207. GE_CHK_STATUS_RET(CalcOpParam(comp_graph), "Graph[%s] builder CalcOpParam() return fail.",
  208. comp_graph->GetName().c_str());
  209. GE_TIMESTAMP_END(CalcOpParam, "GraphBuilder::CalcOpParam");
  210. GE_DUMP(comp_graph, "AfterCalcOpParam");
  211. Graph2SubGraphInfoList subgraph_map;
  212. ge::ModelBuilder builder(comp_graph, subgraph_map, stream_max_parallel_num_, hcom_parallel_, build_mode_);
  213. ModelPtr model_ptr = MakeShared<ge::Model>();
  214. if (model_ptr == nullptr) {
  215. return MEMALLOC_FAILED;
  216. }
  217. GE_TIMESTAMP_START(BuildModelForGetDynShapeTask);
  218. GE_CHK_STATUS_RET(builder.BuildModelForGetDynShapeTask(*model_ptr),
  219. "Graph[%s] builder BuildModelForGetDynShapeTask() return fail.", comp_graph->GetName().c_str());
  220. GE_TIMESTAMP_END(BuildModelForGetDynShapeTask, "GraphBuilder::BuildModelForGetDynShapeTask");
  221. GE_TIMESTAMP_START(GetTaskInfo);
  222. Status ret = GetTaskInfo(builder, model_ptr, comp_graph, subgraph_map, session_id);
  223. GE_TIMESTAMP_END(GetTaskInfo, "GraphBuilder::GetTaskInfo");
  224. GraphUtils::DumpGEGraph(comp_graph, "AfterGetTask");
  225. GraphUtils::DumpGEGraphToOnnx(*comp_graph, "AfterGetTask");
  226. if (ret != SUCCESS) {
  227. GELOGE(ret, "Graph[%s] builder GetTaskInfo() return fail.", comp_graph->GetName().c_str());
  228. return ret;
  229. }
  230. ge_model_ptr = MakeShared<ge::GeModel>();
  231. if (ge_model_ptr == nullptr) {
  232. return MEMALLOC_FAILED;
  233. }
  234. GE_CHK_STATUS_RET(builder.SaveDataToModel(*model_ptr, *ge_model_ptr),
  235. "Graph[%s] builder SaveDataToModel() return fail.", comp_graph->GetName().c_str());
  236. GELOGI("Success to build graph[%s] model.", comp_graph->GetName().c_str());
  237. return SUCCESS;
  238. }
  239. Status GraphBuilder::BuildForDynamicShapeGraph(ComputeGraphPtr &comp_graph,
  240. std::vector<SubGraphInfoPtr> &subgraph_ptr_list,
  241. GeRootModelPtr &ge_root_model_ptr, GeModelPtr &ge_model_ptr,
  242. uint64_t session_id) {
  243. GELOGI("Start to build BuildForDynamicShape for dynamic shape.");
  244. for (const auto &node : comp_graph->GetDirectNode()) {
  245. auto op_desc = node->GetOpDesc();
  246. GE_CHECK_NOTNULL(op_desc);
  247. if (node->GetType() == DATA) {
  248. GE_CHK_STATUS_RET(CalcDynShapeRootGraphDataSize(op_desc), "Calc dynamic shape root graph data[%s] size failed.",
  249. op_desc->GetName().c_str());
  250. }
  251. // ATTR_NAME_IS_UNKNOWN_SHAPE is set on "graph partion" stage, but afer fusion , the graph may
  252. // be changed so here need to renew. For example , the scene followed:
  253. // (known)partioncall(known) (known)partioncall(known)
  254. // After fusion
  255. // | -->
  256. // (known)Unique(unknown)--->(unknow)Shape(unknown) (known)FuncDef(known)
  257. // if scene like this , it should be process as known shape graph
  258. bool is_unknown_shape = false;
  259. GE_CHK_STATUS_RET(ge::NodeUtils::GetNodeUnknownShapeStatus(*node, is_unknown_shape),
  260. "Get node[%s] shape status failed!", node->GetName().c_str());
  261. if (!is_unknown_shape) {
  262. GE_CHK_BOOL_EXEC(ge::AttrUtils::SetBool(op_desc, ATTR_NAME_IS_UNKNOWN_SHAPE, is_unknown_shape), return FAILED,
  263. "Renew node [%s] attr[%s] failed!", node->GetName().c_str(), ATTR_NAME_IS_UNKNOWN_SHAPE.c_str());
  264. GELOGD("renew node [%s] attr[%s] success! value is %d", node->GetName().c_str(),
  265. ATTR_NAME_IS_UNKNOWN_SHAPE.c_str(), is_unknown_shape);
  266. }
  267. vector<string> subgraph_names = op_desc->GetSubgraphInstanceNames();
  268. for (auto subgraph_name : subgraph_names) {
  269. ComputeGraphPtr subgraph = comp_graph->GetSubgraph(subgraph_name);
  270. bool is_unknown_shape = false;
  271. if (!AttrUtils::GetBool(op_desc, ATTR_NAME_IS_UNKNOWN_SHAPE, is_unknown_shape)) {
  272. GELOGE(PARAM_INVALID, "Get op %s unknown shape attr failed.", op_desc->GetName().c_str());
  273. return PARAM_INVALID;
  274. }
  275. if (is_unknown_shape) {
  276. // unknown shape build flow
  277. GE_CHK_STATUS_RET(BuildForUnknownShapeGraph(subgraph, ge_model_ptr, session_id),
  278. "Build for unknown shape graph failed.");
  279. } else {
  280. // known shape build flow
  281. GE_CHK_STATUS_RET(BuildForKnownShapeGraph(subgraph, subgraph_ptr_list, ge_model_ptr, session_id),
  282. "Build for known shape graph failed.");
  283. }
  284. ge_root_model_ptr->SetSubgraphInstanceNameToModel(subgraph_name, ge_model_ptr);
  285. }
  286. }
  287. return SUCCESS;
  288. }
  289. Status GraphBuilder::GetTaskInfo(const ge::ModelBuilder &builder, const ModelPtr &model_ptr,
  290. ComputeGraphPtr &comp_graph, Graph2SubGraphInfoList &subgraph_map,
  291. uint64_t session_id) {
  292. GE_CHECK_NOTNULL(model_ptr);
  293. GE_CHECK_NOTNULL(comp_graph);
  294. int64_t memory_size = 0;
  295. if (!AttrUtils::GetInt(model_ptr, ATTR_MODEL_MEMORY_SIZE, memory_size)) {
  296. GELOGE(INTERNAL_ERROR, "Get memory size fail.");
  297. return INTERNAL_ERROR;
  298. }
  299. int64_t weight_size = 0;
  300. if (!AttrUtils::GetInt(model_ptr, ATTR_MODEL_WEIGHT_SIZE, weight_size)) {
  301. GELOGE(INTERNAL_ERROR, "Get weight memory size fail.");
  302. return INTERNAL_ERROR;
  303. }
  304. auto *get_mem_base =
  305. reinterpret_cast<uint8_t *>(reinterpret_cast<uintptr_t>(ge::VarManager::Instance(0)->GetVarMemMaxSize()));
  306. uint8_t *get_weight_mem_base = get_mem_base;
  307. if (weight_size > 0) {
  308. get_weight_mem_base = get_mem_base + memory_size;
  309. }
  310. RunContextUtil run_context;
  311. Status ret = run_context.InitMemInfo(get_mem_base, memory_size, get_weight_mem_base, weight_size);
  312. if (ret != SUCCESS) {
  313. GELOGE(ret, "task_generator init mem info fail.");
  314. return ret;
  315. }
  316. auto weight_buffer = builder.GetWeightBuffer();
  317. ret = run_context.CreateRunContext(*model_ptr, comp_graph, weight_buffer, session_id);
  318. if (ret != SUCCESS) {
  319. GELOGE(ret, "runContext create run context fail.");
  320. return ret;
  321. }
  322. StreamGraphOptimizer stream_optimizer;
  323. ret = stream_optimizer.OptimizeStreamedSubGraph(comp_graph, subgraph_map, run_context.GetRunContext());
  324. if (ret != SUCCESS) {
  325. GELOGE(ret, "Optimize streamed subGraph fail.");
  326. return ret;
  327. }
  328. GE_DUMP(comp_graph, "AfterOptimizeStreamedSubGraph");
  329. auto *get_var_mem_base =
  330. reinterpret_cast<uint8_t *>(reinterpret_cast<uintptr_t>(ge::VarManager::Instance(0)->GetVarMemLogicBase()));
  331. uint64_t var_size = (ge::VarManager::Instance(session_id)->GetVarMemSize(RT_MEMORY_HBM) > 0)
  332. ? ge::VarManager::Instance(0)->GetVarMemMaxSize()
  333. : 0;
  334. TaskGenerator task_generator(get_var_mem_base, var_size);
  335. ret = task_generator.GetTaskInfo(*model_ptr, comp_graph, session_id, run_context.GetRunContext());
  336. return ret;
  337. }
  338. Status GraphBuilder::SetInputSize(const ge::NodePtr &node_ptr) {
  339. // set input_desc.size = src_node.output_desc.size
  340. if (node_ptr->GetType() == DATA) {
  341. if (UpdateDataInputSize(node_ptr) != SUCCESS) {
  342. GELOGE(FAILED, "Update data input size failed.");
  343. return FAILED;
  344. }
  345. }
  346. for (const auto &in_data_anchor : node_ptr->GetAllInDataAnchors()) {
  347. const auto &peer_out_anchor = in_data_anchor->GetPeerOutAnchor();
  348. GE_IF_BOOL_EXEC(peer_out_anchor == nullptr, continue);
  349. const auto &src_node = peer_out_anchor->GetOwnerNode();
  350. const auto &src_op = src_node->GetOpDesc();
  351. GE_IF_BOOL_EXEC(src_op == nullptr, continue);
  352. auto node_op_desc = node_ptr->GetOpDesc();
  353. GE_IF_BOOL_EXEC(node_op_desc == nullptr, continue);
  354. // set dst_node.input_desc = src_node.output_desc
  355. ge::GeTensorDesc desc_temp(src_op->GetOutputDesc(peer_out_anchor->GetIdx()));
  356. int64_t size = 0;
  357. GE_IF_BOOL_EXEC(ge::TensorUtils::GetSize(desc_temp, size) != SUCCESS, GELOGI("Get size failed!"));
  358. GELOGD("src node %s output desc, dim_size: %zu, mem_size: %ld, format: %s, type: %s.", src_node->GetName().c_str(),
  359. desc_temp.GetShape().GetDimNum(), size, TypeUtils::FormatToSerialString(desc_temp.GetFormat()).c_str(),
  360. TypeUtils::DataTypeToSerialString(desc_temp.GetDataType()).c_str());
  361. for (size_t i = 0; i < desc_temp.GetShape().GetDimNum(); ++i) {
  362. GELOGD("dims[%zu]: %ld", i, desc_temp.GetShape().GetDim(i));
  363. }
  364. auto input_desc = node_op_desc->GetInputDescPtr(in_data_anchor->GetIdx());
  365. GE_CHECK_NOTNULL(input_desc);
  366. ge::TensorUtils::SetSize(const_cast<GeTensorDesc &>(*input_desc), size);
  367. GE_CHK_STATUS_RET(node_op_desc->UpdateInputDesc(in_data_anchor->GetIdx(), *input_desc));
  368. GELOGD("%s input desc, dim_size: %zu, mem_size: %u, format: %s, type: %s.", node_ptr->GetName().c_str(),
  369. input_desc->GetShape().GetDimNum(), size, TypeUtils::FormatToSerialString(input_desc->GetFormat()).c_str(),
  370. TypeUtils::DataTypeToSerialString(input_desc->GetDataType()).c_str());
  371. }
  372. return SUCCESS;
  373. }
  374. Status GraphBuilder::UpdateDataInputSize(const ge::NodePtr &node_ptr) {
  375. const auto &op_desc = node_ptr->GetOpDesc();
  376. if (op_desc == nullptr) {
  377. GELOGE(FAILED, "Op desc is nullptr.");
  378. return FAILED;
  379. }
  380. // data op only has one output anchor
  381. ge::GeTensorDesc output_desc = op_desc->GetOutputDesc(0);
  382. int64_t output_size = 0;
  383. if (ge::TensorUtils::GetSize(output_desc, output_size) != SUCCESS) {
  384. GELOGW("Get size failed!");
  385. }
  386. if (output_size > 0) {
  387. GELOGI("No need to update data input size.");
  388. return SUCCESS;
  389. } else {
  390. int64_t real_dim_size = 0;
  391. ge::graphStatus graph_status = TensorUtils::GetTensorSizeInBytes(output_desc, real_dim_size);
  392. if (graph_status != GRAPH_SUCCESS) {
  393. GELOGE(FAILED, "Get tensor size in bytes failed.");
  394. return FAILED;
  395. }
  396. // data op only has one input anchor
  397. ge::GeTensorDesc input_desc = op_desc->GetInputDesc(0);
  398. ge::TensorUtils::SetSize(input_desc, real_dim_size);
  399. if (op_desc->UpdateInputDesc(0, input_desc) != GRAPH_SUCCESS) {
  400. GELOGE(FAILED, "Update input desc size failed.");
  401. return FAILED;
  402. }
  403. }
  404. return SUCCESS;
  405. }
  406. Status GraphBuilder::CalcDynShapeRootGraphDataSize(const ge::OpDescPtr &op_desc) {
  407. GELOGI("Begin to calc dynamic shape graph data[%s] size.", op_desc->GetName().c_str());
  408. // data op only has one output anchor
  409. ge::GeTensorDesc output_desc = op_desc->GetOutputDesc(0);
  410. int64_t output_size = 0;
  411. if (ge::TensorUtils::GetSize(output_desc, output_size) != SUCCESS) {
  412. GELOGW("Get size failed!");
  413. }
  414. if (output_size > 0) {
  415. GELOGI("No need to update dynamic shape graph data output size[%ld].", output_size);
  416. return SUCCESS;
  417. } else {
  418. int64_t real_dim_size = 0;
  419. ge::graphStatus graph_status = TensorUtils::GetTensorSizeInBytes(output_desc, real_dim_size);
  420. if (graph_status != GRAPH_SUCCESS) {
  421. GELOGE(FAILED, "Get tensor size in bytes failed.");
  422. return FAILED;
  423. }
  424. ge::TensorUtils::SetSize(output_desc, real_dim_size);
  425. GELOGI("Update dynamic shape graph data output size to [%ld].", real_dim_size);
  426. if (op_desc->UpdateOutputDesc(0, output_desc) != GRAPH_SUCCESS) {
  427. GELOGE(FAILED, "Update dynamic shape graph data output desc size failed.");
  428. return FAILED;
  429. }
  430. }
  431. return SUCCESS;
  432. }
  433. Status GraphBuilder::SecondPartition(ge::ComputeGraphPtr &comp_graph, vector<ge::SubGraphInfoPtr> &subgraph_ptr_list) {
  434. GELOGI("[SecondPartition] second partition.");
  435. GE_TIMESTAMP_START(GraphPartition2);
  436. auto ret = graph_partitioner_.Partition(comp_graph, GraphPartitioner::kSecondPartitioning);
  437. if (ret != SUCCESS) {
  438. GELOGE(ret, "Graph partition Failed");
  439. return ret;
  440. }
  441. GE_CHK_STATUS_RET(ret, "Graph partition Failed.");
  442. auto graph_2_subgraphlist = graph_partitioner_.GetSubGraphMap();
  443. if (graph_2_subgraphlist.find(comp_graph) != graph_2_subgraphlist.end()) {
  444. subgraph_ptr_list = graph_2_subgraphlist[comp_graph];
  445. } else {
  446. GELOGE(FAILED, "Find subgraph failed.");
  447. return FAILED;
  448. }
  449. GE_TIMESTAMP_END(GraphPartition2, "GraphPartitioner::Partition2");
  450. return ret;
  451. }
  452. } // namespace ge

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示