You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

ge_executor.cc 40 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119
  1. /**
  2. * Copyright 2019-2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "executor/ge_executor.h"
  17. #include <cce/cce.h>
  18. #include <cce/compiler_stub.h>
  19. #include <ctime>
  20. #include <iostream>
  21. #include "common/debug/log.h"
  22. #include "common/ge/ge_util.h"
  23. #include "common/helper/model_helper.h"
  24. #include "common/profiling/profiling_manager.h"
  25. #include "common/dump/dump_manager.h"
  26. #include "common/util.h"
  27. #include "framework/common/debug/ge_log.h"
  28. #include "framework/common/util.h"
  29. #include "graph/execute/graph_execute.h"
  30. #include "graph/load/graph_loader.h"
  31. #include "graph/load/new_model_manager/davinci_model_parser.h"
  32. #include "graph/load/new_model_manager/model_manager.h"
  33. #include "graph/manager/graph_mem_allocator.h"
  34. #include "graph/model.h"
  35. #include "graph/utils/graph_utils.h"
  36. #include "mmpa/mmpa_api.h"
  37. #include "single_op/single_op_manager.h"
  38. #include "graph/manager/graph_var_manager.h"
  39. #include "graph/load/new_model_manager/davinci_model.h"
  40. using std::string;
  41. using std::vector;
  42. namespace {
  43. const size_t kDynamicBatchSizeVecSize = 1;
  44. const size_t kStaticBatchInfoSize = 1;
  45. const size_t kDynamicImageSizeVecSize = 2;
  46. const size_t kDynamicImageSizeInputSize = 2;
  47. const char *const kBatchLabel = "Batch_";
  48. ge::Status TransferDomiErrorCode(const uint32_t errorCode) {
  49. switch (errorCode) {
  50. case ge::PARAM_INVALID:
  51. case domi::PARAM_INVALID:
  52. return ge::PARAM_INVALID;
  53. case ge::INTERNAL_ERROR:
  54. case domi::INTERNAL_ERROR:
  55. return ge::INTERNAL_ERROR;
  56. default:
  57. return ge::FAILED;
  58. }
  59. }
  60. void GetGeTensorDescFromDomiInfo(std::vector<ge::TensorDesc> &ge_descs,
  61. const std::vector<ge::InputOutputDescInfo> &domi_descs,
  62. const std::vector<uint32_t> &formats) {
  63. uint32_t idx = 0;
  64. for (auto desc_item : domi_descs) {
  65. ge::TensorDesc ge_desc;
  66. ge_desc.SetName(desc_item.name);
  67. ge_desc.SetDataType(static_cast<ge::DataType>(desc_item.data_type));
  68. ge_desc.SetFormat(static_cast<ge::Format>(formats[idx]));
  69. std::vector<int64_t> shape_dims;
  70. for (auto dim : desc_item.shape_info.dims) {
  71. shape_dims.push_back(dim);
  72. }
  73. ge::Shape ge_shape(shape_dims);
  74. ge_desc.SetShape(ge_shape);
  75. ge_desc.SetSize(desc_item.size);
  76. ge_descs.emplace_back(ge_desc);
  77. ++idx;
  78. }
  79. }
  80. void GetDomiInputData(const ge::RunModelData &input_data, ge::InputData &inputs) {
  81. inputs.index = input_data.index;
  82. inputs.model_id = input_data.modelId;
  83. inputs.timestamp = input_data.timestamp;
  84. inputs.timeout = input_data.timeout;
  85. inputs.request_id = input_data.request_id;
  86. for (const auto &data_item : input_data.blobs) {
  87. ge::DataBuffer dataBuf{data_item.data, data_item.length, data_item.isDataSupportMemShare};
  88. inputs.blobs.emplace_back(dataBuf);
  89. }
  90. }
  91. void GetDomiOutputData(const ge::RunModelData &output_data, ge::OutputData &outputs) {
  92. outputs.index = output_data.index;
  93. outputs.model_id = output_data.modelId;
  94. for (const auto &data_item : output_data.blobs) {
  95. ge::DataBuffer dataBuf(data_item.data, data_item.length, data_item.isDataSupportMemShare);
  96. outputs.blobs.emplace_back(dataBuf);
  97. }
  98. }
  99. void SetDynamicInputDataFlag(const ge::RunModelData &input_data, const std::vector<std::vector<int64_t>> batch_info,
  100. ge::InputData &inputs) {
  101. inputs.is_dynamic_batch = true;
  102. std::string batch_label;
  103. size_t match_idx = 0;
  104. for (size_t i = 0; i < batch_info.size(); ++i) {
  105. // dynamic_dims
  106. if (input_data.dynamic_dims.size() != 0) {
  107. bool is_match = true;
  108. for (size_t j = 0; j < static_cast<size_t>(input_data.dynamic_dims.size()); ++j) {
  109. if (static_cast<uint64_t>(batch_info[i][j]) != input_data.dynamic_dims[j]) {
  110. is_match = false;
  111. break;
  112. }
  113. }
  114. if (is_match) {
  115. match_idx = i;
  116. break;
  117. }
  118. // dynamic_batch_size
  119. } else if (batch_info[i].size() == kDynamicBatchSizeVecSize &&
  120. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_batch_size)) {
  121. match_idx = i;
  122. break;
  123. // dynamic_image_size
  124. } else if (batch_info[i].size() == kDynamicImageSizeVecSize &&
  125. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_image_height) &&
  126. batch_info[i][1] == static_cast<int64_t>(input_data.dynamic_image_width)) {
  127. match_idx = i;
  128. break;
  129. }
  130. }
  131. batch_label = kBatchLabel + std::to_string(match_idx);
  132. inputs.batch_label = batch_label;
  133. GELOGI("current batch label:%s", batch_label.c_str());
  134. }
  135. bool IsDynamicBatchSizeMatchModel(uint64_t batch_size, const vector<std::vector<int64_t>> &batch_info) {
  136. if (batch_info.empty()) {
  137. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  138. return false;
  139. }
  140. for (auto batch : batch_info) {
  141. if (batch.size() != kDynamicBatchSizeVecSize) {
  142. GELOGE(ge::FAILED, "Dynamic batch param num is %zu, current batch size is %zu.", kDynamicBatchSizeVecSize,
  143. batch.size());
  144. return false;
  145. }
  146. if (batch[0] == static_cast<int64_t>(batch_size)) {
  147. return true;
  148. }
  149. }
  150. GELOGE(ge::FAILED, "Dynamic batch %lu can not match the gear of model.", batch_size);
  151. return false;
  152. }
  153. bool IsDynamicImageSizeMatchModel(uint64_t image_height, uint64_t image_width,
  154. const vector<std::vector<int64_t>> &batch_info) {
  155. if (batch_info.empty()) {
  156. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  157. return false;
  158. }
  159. for (auto resolution : batch_info) {
  160. if (resolution.size() != kDynamicImageSizeVecSize) {
  161. GELOGE(ge::FAILED, "Dynamic resolution param num is %zu, current resolution size is %zu.",
  162. kDynamicImageSizeVecSize, resolution.size());
  163. return false;
  164. }
  165. if (resolution[0] == static_cast<int64_t>(image_height) && resolution[1] == static_cast<int64_t>(image_width)) {
  166. return true;
  167. }
  168. }
  169. GELOGE(ge::FAILED, "Dynamic resolution (%lu,%lu) can not match the gear of model.", image_height, image_width);
  170. return false;
  171. }
  172. bool IsDynmaicDimsSizeMatchModel(const vector<uint64_t> cur_dynamic_dims, const vector<vector<int64_t>> &batch_info) {
  173. if (batch_info.empty()) {
  174. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  175. return false;
  176. }
  177. bool find_match = false;
  178. for (auto resolution : batch_info) {
  179. if (cur_dynamic_dims.size() != resolution.size()) {
  180. GELOGE(ge::FAILED, "Cur dynamic dims param num is %zu, current resolution size is %zu.", cur_dynamic_dims.size(),
  181. resolution.size());
  182. return false;
  183. }
  184. bool flag = true;
  185. for (std::size_t i = 0; i < resolution.size(); ++i) {
  186. if (cur_dynamic_dims[i] != static_cast<uint64_t>(resolution[i])) {
  187. flag = false;
  188. break;
  189. }
  190. }
  191. if (flag) {
  192. find_match = true;
  193. break;
  194. }
  195. }
  196. if (!find_match) {
  197. GELOGE(ge::FAILED, "choose dynamic dims can not match the gear of model.");
  198. }
  199. return find_match;
  200. }
  201. } // namespace
  202. namespace ge {
  203. bool GeExecutor::isInit_ = false;
  204. class ModelListenerAdapter : public ModelListener {
  205. public:
  206. domi::Status OnComputeDone(uint32_t model_id, uint32_t dataIndex, uint32_t resultCode,
  207. std::vector<ge::OutputTensorInfo> &outputs) {
  208. if (listener == nullptr) {
  209. GELOGE(ge::FAILED, "listener is null.");
  210. return FAILED;
  211. }
  212. return listener->OnComputeDone(model_id, dataIndex, resultCode, outputs);
  213. }
  214. std::shared_ptr<ge::ModelListener> listener;
  215. };
  216. GeExecutor::GeExecutor() {}
  217. Status GeExecutor::Initialize() {
  218. GELOGI("Init GeExecutor begin.");
  219. if (isInit_) {
  220. GELOGW("Already initialized, no need to be initialized again.");
  221. return ge::SUCCESS;
  222. }
  223. std::vector<rtMemType_t> mem_type(1, RT_MEMORY_HBM);
  224. auto ret = MemManager::Instance().Initialize(mem_type);
  225. if (ret != SUCCESS) {
  226. GELOGE(ret, "Memory Manager init failed.");
  227. return ret;
  228. }
  229. // Start profiling
  230. Options profiling_options;
  231. profiling_options.device_id = 0;
  232. profiling_options.job_id = "";
  233. ProfilingManager::Instance().Init(profiling_options);
  234. isInit_ = true;
  235. GELOGI("Init GeExecutor over.");
  236. return ge::SUCCESS;
  237. }
  238. Status GeExecutor::Finalize() {
  239. GELOGI("Uninit GeExecutor begin.");
  240. if (isInit_ == false) {
  241. GELOGW("GeExecutor has not been initialized.");
  242. return ge::SUCCESS;
  243. }
  244. // Stop profiling
  245. if (ProfilingManager::Instance().ProfilingOn()) {
  246. ProfilingManager::Instance().StopProfiling();
  247. ProfilingManager::Instance().PluginUnInit(GE_PROFILING_MODULE);
  248. }
  249. GELOGI("Uninit GeExecutor over.");
  250. return ge::SUCCESS;
  251. }
  252. Status GeExecutor::SetDynamicBatchSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  253. uint64_t batch_size) {
  254. if (dynamic_input_addr == nullptr) {
  255. GELOGE(PARAM_INVALID, "Dynamic input addr is nullptr!");
  256. return PARAM_INVALID;
  257. }
  258. uint64_t size = sizeof(uint64_t);
  259. if (length < size) {
  260. GELOGE(PARAM_INVALID, "Dynamic input size [%lu] is less than [%lu]!", length, size);
  261. return PARAM_INVALID;
  262. }
  263. // Verify whether the input dynamic batch matches the model gear
  264. std::vector<std::vector<int64_t>> batch_info;
  265. std::vector<uint64_t> batch_num{batch_size};
  266. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  267. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  268. if (ret != SUCCESS) {
  269. GELOGE(ret, "Get dynamic input info failed.");
  270. return ret;
  271. }
  272. if (!IsDynamicBatchSizeMatchModel(batch_size, batch_info)) {
  273. GELOGE(PARAM_INVALID, "The current dynamic input does not match the gear of the model.");
  274. return PARAM_INVALID;
  275. }
  276. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_BATCH));
  277. if (ret != SUCCESS) {
  278. GELOGE(ret, "Set dynamic size failed");
  279. return ret;
  280. }
  281. // memcpy dynamic_batch_size from host to device
  282. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &batch_size, size, RT_MEMCPY_HOST_TO_DEVICE);
  283. if (rt_ret != RT_ERROR_NONE) {
  284. GELOGE(RT_FAILED, "memcpy dynamic batch input data failed! ret: 0x%X", rt_ret);
  285. return RT_ERROR_TO_GE_STATUS(rt_ret);
  286. }
  287. return SUCCESS;
  288. }
  289. Status GeExecutor::SetDynamicImageSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  290. uint64_t image_height, uint64_t image_width) {
  291. if (dynamic_input_addr == nullptr) {
  292. GELOGE(PARAM_INVALID, "Dynamic input addr is nullptr!");
  293. return PARAM_INVALID;
  294. }
  295. uint64_t dynamic_input_size = kDynamicImageSizeInputSize * sizeof(uint64_t);
  296. if (length < dynamic_input_size) {
  297. GELOGE(PARAM_INVALID, "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  298. return PARAM_INVALID;
  299. }
  300. // Verify whether the input dynamic resolution matches the model gear
  301. std::vector<std::vector<int64_t>> batch_info;
  302. std::vector<uint64_t> batch_num{image_height, image_width};
  303. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  304. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  305. if (ret != SUCCESS) {
  306. GELOGE(ret, "Get dynamic input info failed.");
  307. return ret;
  308. }
  309. if (!IsDynamicImageSizeMatchModel(image_height, image_width, batch_info)) {
  310. GELOGE(PARAM_INVALID, "The current dynamic input does not match the gear of the model.");
  311. return PARAM_INVALID;
  312. }
  313. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_IMAGE));
  314. if (ret != SUCCESS) {
  315. GELOGE(ret, "Set dynamic size failed");
  316. return ret;
  317. }
  318. // Memcpy dynamic resolution height from host to device
  319. rtError_t rt_ret =
  320. rtMemcpy(dynamic_input_addr, sizeof(uint64_t), &image_height, sizeof(uint64_t), RT_MEMCPY_HOST_TO_DEVICE);
  321. if (rt_ret != RT_ERROR_NONE) {
  322. GELOGE(RT_FAILED, "memcpy dynamic resolution input data failed! ret: 0x%X", rt_ret);
  323. return RT_ERROR_TO_GE_STATUS(rt_ret);
  324. }
  325. uint64_t remain_size = length - sizeof(uint64_t);
  326. // Memcpy dynamic resolution width from host to device
  327. if (rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + sizeof(uint64_t)),
  328. remain_size, &image_width, sizeof(uint64_t), RT_MEMCPY_HOST_TO_DEVICE) != RT_ERROR_NONE) {
  329. GELOGE(FAILED, "memcpy dynamic resolution input data failed!");
  330. return FAILED;
  331. }
  332. return SUCCESS;
  333. }
  334. Status GeExecutor::SetDynamicDims(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  335. const vector<uint64_t> &dynamic_dims) {
  336. if (dynamic_input_addr == nullptr) {
  337. GELOGE(FAILED, "Dynamic input addr is nullptr!");
  338. return FAILED;
  339. }
  340. vector<uint64_t> cur_dynamic_dims;
  341. Status ret = GetCurDynamicDims(model_id, dynamic_dims, cur_dynamic_dims);
  342. if (ret != SUCCESS) {
  343. GELOGE(FAILED, "Set cur gear dynamic dims failed");
  344. return FAILED;
  345. }
  346. std::vector<std::vector<int64_t>> batch_info;
  347. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  348. ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  349. if (ret != SUCCESS) {
  350. GELOGE(ret, "Get dynamic input info failed.");
  351. return ret;
  352. }
  353. if (!IsDynmaicDimsSizeMatchModel(cur_dynamic_dims, batch_info)) {
  354. GELOGE(PARAM_INVALID, "The current dynamic input does not match the gear of the model.");
  355. return PARAM_INVALID;
  356. }
  357. ret = GraphExecutor::SetDynamicSize(model_id, cur_dynamic_dims, static_cast<int32_t>(DYNAMIC_DIMS));
  358. if (ret != SUCCESS) {
  359. GELOGE(FAILED, "Set dynamic size failed");
  360. return FAILED;
  361. }
  362. size_t dynamic_dim_num = cur_dynamic_dims.size();
  363. uint64_t dynamic_input_size = static_cast<uint64_t>(dynamic_dim_num * sizeof(uint64_t));
  364. if (length < dynamic_input_size) {
  365. GELOGE(FAILED, "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  366. return FAILED;
  367. }
  368. for (uint32_t i = 0; i < dynamic_dim_num; ++i) {
  369. // Memcpy dynamic dim[i] from host to device
  370. if (rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + sizeof(uint64_t) * i),
  371. length - sizeof(uint64_t) * i, &cur_dynamic_dims[i], sizeof(uint64_t),
  372. RT_MEMCPY_HOST_TO_DEVICE) != RT_ERROR_NONE) {
  373. GELOGE(FAILED, "memcpy dynamic resolution input data failed!");
  374. return FAILED;
  375. }
  376. }
  377. return SUCCESS;
  378. }
  379. Status GeExecutor::GetCurDynamicDims(uint32_t model_id, const vector<uint64_t> &dynamic_dims,
  380. vector<uint64_t> &cur_dynamic_dims) {
  381. cur_dynamic_dims.clear();
  382. vector<ge::TensorDesc> input_desc;
  383. vector<ge::TensorDesc> output_desc;
  384. auto ret = GetModelDescInfo(model_id, input_desc, output_desc);
  385. if (ret != ge::SUCCESS) {
  386. GELOGE(FAILED, "GetModelDescInfo failed.");
  387. return FAILED;
  388. }
  389. vector<string> user_designate_shape_order;
  390. vector<int64_t> all_data_dims;
  391. ret = GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  392. if (ret != ge::SUCCESS) {
  393. GELOGE(FAILED, "GetUserDesignateShapeOrder failed.");
  394. return FAILED;
  395. }
  396. for (auto &data_name : user_designate_shape_order) {
  397. for (auto &desc : input_desc) {
  398. if (desc.GetName() == data_name) {
  399. for (auto dim : desc.GetShape().GetDims()) {
  400. all_data_dims.push_back(dim);
  401. }
  402. break;
  403. }
  404. }
  405. }
  406. if (dynamic_dims.size() != all_data_dims.size()) {
  407. GELOGE(FAILED, "Dynamic input size [%lu] is not equal with all data dims size [%lu]!", dynamic_dims.size(),
  408. all_data_dims.size());
  409. return FAILED;
  410. }
  411. for (std::size_t i = 0; i < all_data_dims.size(); ++i) {
  412. if (all_data_dims[i] < 0) {
  413. cur_dynamic_dims.push_back(dynamic_dims[i]);
  414. } else if (static_cast<uint64_t>(all_data_dims[i]) != dynamic_dims[i]) {
  415. GELOGE(PARAM_INVALID, "Static dims should be same, index: %zu value: %d should be %d", i, dynamic_dims[i],
  416. all_data_dims[i]);
  417. return PARAM_INVALID;
  418. }
  419. }
  420. return SUCCESS;
  421. }
  422. Status GeExecutor::GetCurShape(const uint32_t model_id, std::vector<int64_t> &batch_info, int32_t &dynamic_type) {
  423. GELOGI("Begin to get current shape");
  424. if (!isInit_) {
  425. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  426. return GE_EXEC_NOT_INIT;
  427. }
  428. Status ret = GraphExecutor::GetCurShape(model_id, batch_info, dynamic_type);
  429. if (ret != SUCCESS) {
  430. GELOGE(ret, "Get current shape failed");
  431. return ret;
  432. }
  433. return SUCCESS;
  434. }
  435. Status GeExecutor::SetDynamicAippData(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  436. const std::vector<kAippDynamicBatchPara> &aippBatchPara,
  437. const kAippDynamicPara &aippParms) {
  438. GELOGI("Enter to SetDynamicAippData.");
  439. if (dynamic_input_addr == nullptr) {
  440. GELOGE(PARAM_INVALID, "Dynamic aipp input addr is nullptr!");
  441. return PARAM_INVALID;
  442. }
  443. if (aippBatchPara.empty()) {
  444. GELOGE(PARAM_INVALID, "aippBatchPara is empty.");
  445. return PARAM_INVALID;
  446. }
  447. uint64_t batch_num = aippBatchPara.size();
  448. uint64_t real_aippParms_size = sizeof(kAippDynamicPara) - sizeof(kAippDynamicBatchPara);
  449. uint64_t struct_len = batch_num * sizeof(kAippDynamicBatchPara) + real_aippParms_size;
  450. GELOGI(
  451. "Get acl input dynamic aipp data, model_id is %u, length is %lu,"
  452. "batch num is %lu, struct_len is %lu",
  453. model_id, length, batch_num, struct_len);
  454. if (struct_len > length) {
  455. GELOGE(PARAM_INVALID, "input dynamic aipp param len [%lu] is larger than aipp_data size [%lu]", struct_len, length);
  456. return PARAM_INVALID;
  457. }
  458. // Memcpy real kAippDynamicBatchPara from host to device
  459. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &aippParms, real_aippParms_size, RT_MEMCPY_HOST_TO_DEVICE);
  460. if (rt_ret != RT_ERROR_NONE) {
  461. GELOGE(RT_FAILED, "memcpy real_aippParms_size failed! ret: 0x%X", rt_ret);
  462. return RT_ERROR_TO_GE_STATUS(rt_ret);
  463. }
  464. uint64_t remain_len = length - real_aippParms_size;
  465. uint8_t *aipp_batch_para_dev = reinterpret_cast<uint8_t *>(dynamic_input_addr) + real_aippParms_size;
  466. for (uint64_t i = 0; i < batch_num; ++i) {
  467. rt_ret = rtMemcpy(reinterpret_cast<void *>(aipp_batch_para_dev + i * sizeof(kAippDynamicBatchPara)),
  468. (remain_len - i * sizeof(kAippDynamicBatchPara)), &(aippBatchPara[i]),
  469. sizeof(kAippDynamicBatchPara), RT_MEMCPY_HOST_TO_DEVICE);
  470. if (rt_ret != RT_ERROR_NONE) {
  471. GELOGE(RT_FAILED, "memcpy kAippDynamicBatchPara input data failed! ret: 0x%X", rt_ret);
  472. return RT_ERROR_TO_GE_STATUS(rt_ret);
  473. }
  474. }
  475. return SUCCESS;
  476. }
  477. // Load model
  478. Status GeExecutor::LoadModelOffline(uint32_t &model_id, const std::string &path, const std::string &key,
  479. int32_t priority, std::shared_ptr<ge::ModelListener> listener) {
  480. GELOGI("load model offline begin.");
  481. if (!isInit_) {
  482. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  483. return GE_EXEC_NOT_INIT;
  484. }
  485. string filePath = RealPath(path.c_str());
  486. if (filePath.empty()) {
  487. GELOGE(ge::FAILED, "File path is invalid. please check your text file '%s'.", path.c_str());
  488. return ge::FAILED;
  489. }
  490. std::shared_ptr<ModelListenerAdapter> listener_adapter = MakeShared<ModelListenerAdapter>();
  491. if (listener_adapter == nullptr) {
  492. GELOGE(MEMALLOC_FAILED, "ModelListenerAdapter make shared failed!");
  493. return ge::FAILED;
  494. }
  495. listener_adapter->listener = listener;
  496. Status ret = GraphLoader::LoadModelFromFile(path, key, priority, listener_adapter, model_id);
  497. if (ret != SUCCESS) {
  498. GELOGE(ret, "[GeExecutor] LoadModelFromFile failed");
  499. return TransferDomiErrorCode(ret);
  500. }
  501. return SUCCESS;
  502. }
  503. Status GeExecutor::LoadModel(uint32_t &model_id, const ModelData &model_data,
  504. std::shared_ptr<ge::ModelListener> listener) {
  505. GELOGI("Load model begin.");
  506. if (!isInit_) {
  507. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  508. return GE_EXEC_NOT_INIT;
  509. }
  510. std::shared_ptr<ModelListenerAdapter> listener_adapter = MakeShared<ModelListenerAdapter>();
  511. if (listener_adapter == nullptr) {
  512. GELOGE(MEMALLOC_FAILED, "ModelListenerAdapter make shared failed!");
  513. return ge::FAILED;
  514. }
  515. listener_adapter->listener = listener;
  516. Status ret = GraphLoader::LoadModel(model_data, listener_adapter, model_id);
  517. if (ret != SUCCESS) {
  518. GELOGE(ret, "[GeExecutor] LoadModel failed.");
  519. return TransferDomiErrorCode(ret);
  520. }
  521. return ret;
  522. }
  523. Status GeExecutor::UnloadModel(uint32_t model_id) {
  524. GELOGI("unload model %u begin.", model_id);
  525. if (!isInit_) {
  526. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  527. return GE_EXEC_NOT_INIT;
  528. }
  529. Status ret = GraphLoader::DestroyAicpuSessionForInfer(model_id);
  530. if (ret != SUCCESS) {
  531. GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
  532. return FAILED;
  533. }
  534. std::shared_ptr<DavinciModel> davinci_model = ModelManager::GetInstance()->GetModel(model_id);
  535. if (davinci_model != nullptr) {
  536. uint64_t session_id = davinci_model->GetSessionId();
  537. VarManagerPool::Instance().RemoveVarManager(session_id);
  538. }
  539. return GraphLoader::UnloadModel(model_id);
  540. }
  541. Status GeExecutor::RunModel(const ge::RunModelData &input_data, ge::RunModelData &output_data) {
  542. GELOGI("run model begin.");
  543. if (!isInit_) {
  544. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  545. return GE_EXEC_NOT_INIT;
  546. }
  547. InputData inputs;
  548. GetDomiInputData(input_data, inputs);
  549. OutputData outputs;
  550. GetDomiOutputData(output_data, outputs);
  551. return GraphExecutor::DataInput(inputs, outputs);
  552. }
  553. // Get input and output descriptor
  554. Status GeExecutor::GetModelDescInfo(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
  555. std::vector<ge::TensorDesc> &output_desc, bool new_model_desc) {
  556. GELOGI("get model desc info begin.");
  557. if (!isInit_) {
  558. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  559. return GE_EXEC_NOT_INIT;
  560. }
  561. std::vector<InputOutputDescInfo> input_desc_infos;
  562. std::vector<InputOutputDescInfo> output_desc_infos;
  563. std::vector<uint32_t> input_formats;
  564. std::vector<uint32_t> output_formats;
  565. Status ret = GraphExecutor::GetInputOutputDescInfo(model_id, input_desc_infos, output_desc_infos, input_formats,
  566. output_formats, new_model_desc);
  567. if (ret != domi::SUCCESS) {
  568. GELOGE(ret, "GetInputOutputDescInfo failed. ret = %u", ret);
  569. return ret;
  570. }
  571. if (input_formats.size() != input_desc_infos.size()) {
  572. GELOGE(ge::PARAM_INVALID, "input_formats size %zu is not equal to input_desc_infos size %zu.", input_formats.size(),
  573. input_desc_infos.size());
  574. return ge::PARAM_INVALID;
  575. }
  576. if (output_formats.size() != output_desc_infos.size()) {
  577. GELOGE(ge::PARAM_INVALID, "output_formats size %zu is not equal to output_desc_infos size %zu.",
  578. output_formats.size(), output_desc_infos.size());
  579. return ge::PARAM_INVALID;
  580. }
  581. // Transfer data to TensorDesc
  582. GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
  583. GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
  584. GELOGI("get model desc info end.");
  585. return ge::SUCCESS;
  586. }
  587. ///
  588. /// @ingroup ge
  589. /// @brief Get dynamic batch_info
  590. /// @param [in] model_id
  591. /// @param [out] batch_info
  592. /// @param [out] dynamic_type
  593. /// @return execute result
  594. ///
  595. Status GeExecutor::GetDynamicBatchInfo(uint32_t model_id, std::vector<std::vector<int64_t>> &batch_info,
  596. int32_t &dynamic_type) {
  597. GELOGI("Begin to get dynamic batch info.");
  598. if (!isInit_) {
  599. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  600. return GE_EXEC_NOT_INIT;
  601. }
  602. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  603. if (ret != SUCCESS) {
  604. GELOGE(ret, "GetDynamicBatchInfo failed.");
  605. return ret;
  606. }
  607. GELOGI("Get dynamic batch info succ.");
  608. return SUCCESS;
  609. }
  610. ///
  611. /// @ingroup ge
  612. /// @brief Get combined dynamic dims info
  613. /// @param [in] model_id
  614. /// @param [out] batch_info
  615. /// @return execute result
  616. ///
  617. Status GeExecutor::GetCombinedDynamicDims(uint32_t model_id, vector<vector<int64_t>> &batch_info) {
  618. GELOGI("Begin to get combined dynamic dims info.");
  619. if (!isInit_) {
  620. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  621. return GE_EXEC_NOT_INIT;
  622. }
  623. Status ret = GraphExecutor::GetCombinedDynamicDims(model_id, batch_info);
  624. if (ret != SUCCESS) {
  625. GELOGE(ret, "GetCombinedDynamicDims failed.");
  626. return ret;
  627. }
  628. GELOGI("Get combined dynamic dims succ.");
  629. return SUCCESS;
  630. }
  631. ///
  632. /// @ingroup ge
  633. /// @brief Get user designeate shape order
  634. /// @param [in] model_id
  635. /// @param [out] user_designate_shape_order
  636. /// @return execute result
  637. ///
  638. Status GeExecutor::GetUserDesignateShapeOrder(uint32_t model_id, vector<string> &user_designate_shape_order) {
  639. GELOGI("Begin to get user designate shape info.");
  640. if (!isInit_) {
  641. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  642. return GE_EXEC_NOT_INIT;
  643. }
  644. Status ret = GraphExecutor::GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  645. if (ret != SUCCESS) {
  646. GELOGE(ret, "GetUserDesignateShapeOrder failed.");
  647. return ret;
  648. }
  649. GELOGI("Get user designate shape order succ.");
  650. return SUCCESS;
  651. }
  652. ///
  653. /// @ingroup ge
  654. /// @brief Get AIPP input format
  655. /// @param [in] model_id
  656. /// @param [in] index
  657. /// @param [out] input_format
  658. /// @return execute result
  659. ///
  660. Status GeExecutor::GetAIPPInfo(uint32_t model_id, uint32_t index, AippConfigInfo &aipp_info) {
  661. GELOGI("Begin to GetAIPPInfo.");
  662. if (!isInit_) {
  663. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  664. return GE_EXEC_NOT_INIT;
  665. }
  666. Status ret = GraphExecutor::GetAIPPInfo(model_id, index, aipp_info);
  667. if (ret != SUCCESS) {
  668. GELOGW("GetAIPPInfo is not success.");
  669. return ret;
  670. }
  671. GELOGI("GetAIPPInfo succ.");
  672. return SUCCESS;
  673. }
  674. Status GeExecutor::GetAippType(uint32_t model_id, uint32_t index, InputAippType &type, size_t &aipp_index) {
  675. GELOGI("Begin to get aipp type.");
  676. if (!isInit_) {
  677. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  678. return GE_EXEC_NOT_INIT;
  679. }
  680. Status ret = GraphExecutor::GetAippType(model_id, index, type, aipp_index);
  681. if (ret != SUCCESS) {
  682. GELOGW("Get aipp type is not success.");
  683. return ret;
  684. }
  685. GELOGI("Get aipp type success.");
  686. return SUCCESS;
  687. }
  688. Status GeExecutor::GetModelAttr(uint32_t model_id, std::vector<std::string> &dynamic_output_shape_info) {
  689. GELOGI("Begin to get dynamic batch output shape info");
  690. if (!isInit_) {
  691. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  692. return GE_EXEC_NOT_INIT;
  693. }
  694. Status ret = GraphExecutor::GetModelAttr(model_id, dynamic_output_shape_info);
  695. if (ret != SUCCESS) {
  696. GELOGE(ret, "Get dynamic batch output shape info failed.");
  697. return ret;
  698. }
  699. GELOGI("Get dynamic batch output shape info succ.");
  700. return SUCCESS;
  701. }
  702. Status GeExecutor::GetModelDescInfoForZeroCopy(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
  703. std::vector<TensorDesc> &output_desc) {
  704. GELOGI("get model desc info for zero copy begin.");
  705. if (!isInit_) {
  706. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  707. return GE_EXEC_NOT_INIT;
  708. }
  709. std::vector<InputOutputDescInfo> input_desc_infos;
  710. std::vector<InputOutputDescInfo> output_desc_infos;
  711. std::vector<uint32_t> input_formats;
  712. std::vector<uint32_t> output_formats;
  713. Status ret = GraphExecutor::GetInputOutputDescInfoForZeroCopy(model_id, input_desc_infos, output_desc_infos,
  714. input_formats, output_formats);
  715. if (ret != domi::SUCCESS) {
  716. GELOGE(ret, "Get DescInfo from zero copy failed. ret = %u", ret);
  717. return TransferDomiErrorCode(ret);
  718. }
  719. if (input_formats.size() != input_desc_infos.size()) {
  720. GELOGE(ge::FAILED, "input_formats.size() != input_desc_infos.size().");
  721. return ge::FAILED;
  722. }
  723. if (output_formats.size() != output_desc_infos.size()) {
  724. GELOGE(ge::FAILED, "output_formats.size() != output_desc_infos.size().");
  725. return ge::FAILED;
  726. }
  727. GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
  728. GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
  729. GELOGI("get model desc info from zero copy end.");
  730. return ge::SUCCESS;
  731. }
  732. Status GeExecutor::CommandHandle(const Command &command) {
  733. GELOGI("command handle begin.");
  734. Status ret = GraphLoader::CommandHandle(command);
  735. if (ret != SUCCESS) {
  736. GELOGE(ret, "CommandHandle: Command Handle failed.");
  737. return TransferDomiErrorCode(ret);
  738. }
  739. return SUCCESS;
  740. }
  741. Status GeExecutor::GetMaxUsedMemory(uint32_t model_id, uint32_t &max_size) {
  742. GELOGI("Get max used memory begin.");
  743. if (!isInit_) {
  744. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  745. return GE_EXEC_NOT_INIT;
  746. }
  747. uint64_t max_mem_size = 0;
  748. Status ret = GraphLoader::GetMaxUsedMemory(model_id, max_mem_size);
  749. max_size = static_cast<uint32_t>(max_mem_size);
  750. return ret;
  751. }
  752. /**
  753. * @ingroup ge
  754. * @brief Load data from model file to memory
  755. * @param [in] const std::string &path: Offline model file path
  756. * @param [out] domi::ModelData &model_data: Offline model memory data
  757. * @return SUCCESS handle successfully / others handle failed
  758. */
  759. Status GeExecutor::LoadDataFromFile(const std::string &path, ModelData &model_data) {
  760. GELOGI("Load data from file begin.");
  761. if (!isInit_) {
  762. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  763. return GE_EXEC_NOT_INIT;
  764. }
  765. string filePath = RealPath(path.c_str());
  766. if (filePath.empty()) {
  767. GELOGE(GE_EXEC_MODEL_PATH_INVALID, "File path is invalid. please check your text file '%s'.", path.c_str());
  768. return GE_EXEC_MODEL_PATH_INVALID;
  769. }
  770. GELOGI("load modelData from file: %s.", path.c_str());
  771. std::string key_path;
  772. int32_t priority = 0;
  773. Status ret = GraphLoader::LoadDataFromFile(path, key_path, priority, model_data);
  774. if (ret != SUCCESS) {
  775. if (model_data.model_data != nullptr) {
  776. delete[] static_cast<char *>(model_data.model_data);
  777. model_data.model_data = nullptr;
  778. }
  779. }
  780. return ret;
  781. }
  782. /**
  783. * @ingroup ge
  784. * @brief Load model from offline model memory data
  785. * @param [in] domi::ModelData &model_data: Offline model data
  786. void *dev_ptr: Input/Output memory start address
  787. size_t memsize: Input/Output memory length
  788. void *weight_ptr: Weight memory start address
  789. size_t weightsize: Weight memory length
  790. * @param [out] uint32_t &model_id: identification after model loading
  791. * @return SUCCESS handle successfully / others handle failed
  792. */
  793. Status GeExecutor::LoadModelFromData(uint32_t &model_id, const ModelData &model_data, void *dev_ptr, size_t mem_size,
  794. void *weight_ptr, size_t weight_size) {
  795. GELOGI("Load model from data begin.");
  796. if (!isInit_) {
  797. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  798. return GE_EXEC_NOT_INIT;
  799. }
  800. return GraphLoader::LoadModelFromData(model_id, model_data, dev_ptr, mem_size, weight_ptr, weight_size);
  801. }
  802. /**
  803. * @ingroup ge
  804. * @brief Load task list from ModelData with queue.
  805. * @param [out] model_id: model id allocate from manager.
  806. * @param [in] ge_model_data: Model data load from offline model.
  807. * @param [in] input_queue_ids: input queue ids create from user.
  808. * @param [in] output_queue_ids: input queue ids create from user.
  809. * @return: 0 for success / others for fail
  810. */
  811. Status GeExecutor::LoadModelWithQ(uint32_t &model_id, const ModelData &model_data,
  812. const std::vector<uint32_t> &input_queue_ids,
  813. const std::vector<uint32_t> &output_queue_ids) {
  814. GELOGI("Load model with queue begin.");
  815. if (!isInit_) {
  816. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  817. return GE_EXEC_NOT_INIT;
  818. }
  819. return GraphLoader::LoadModelWithQ(model_id, model_data, input_queue_ids, output_queue_ids);
  820. }
  821. /**
  822. * @ingroup ge
  823. * @brief Synchronous execution of offline model(Do not create thread)
  824. * @param [in] uint32_t model_id: Model ID to execute
  825. void* stream: stream to execute
  826. const domi::InputData *input_data: Model input data
  827. bool async_mode: is asynchronize mode.
  828. * @param [out] domi::OutputData *output_data: Model output data
  829. * @return SUCCESS handle successfully / others handle failed
  830. */
  831. Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
  832. ge::RunModelData &run_output_data, bool async_mode) {
  833. GELOGI("Execute model begin.");
  834. if (!isInit_) {
  835. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  836. return GE_EXEC_NOT_INIT;
  837. }
  838. InputData input_data;
  839. OutputData output_data;
  840. GetDomiInputData(run_input_data, input_data);
  841. GetDomiOutputData(run_output_data, output_data);
  842. if ((run_input_data.dynamic_batch_size != 0) || (run_input_data.dynamic_image_width != 0) ||
  843. (run_input_data.dynamic_image_height != 0) || (run_input_data.dynamic_dims.size() != 0)) {
  844. std::vector<std::vector<int64_t>> batch_info;
  845. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  846. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  847. if (ret != SUCCESS) {
  848. GELOGE(ret, "Get dynamic input info failed.");
  849. return ret;
  850. }
  851. if (!batch_info.empty()) {
  852. SetDynamicInputDataFlag(run_input_data, batch_info, input_data);
  853. }
  854. }
  855. return GraphLoader::ExecuteModel(model_id, stream, async_mode, input_data, output_data);
  856. }
  857. /**
  858. * @ingroup ge
  859. * @brief Get weight memory size from model file
  860. * @param [in] const std::string &path: Offline model file path
  861. * @param [out] size_t &mem_size Execution memory size
  862. size_t &weight_size Weight memory space size
  863. * @return SUCCESS handle successfully / others handle failed
  864. */
  865. Status GeExecutor::GetMemAndWeightSize(const std::string &path, size_t &mem_size, size_t &weight_size) {
  866. GELOGI("Get memory and weight size from file begin.");
  867. if (!isInit_) {
  868. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  869. return GE_EXEC_NOT_INIT;
  870. }
  871. ModelData model;
  872. std::string key;
  873. Status ret = ge::GraphLoader::LoadDataFromFile(path, key, 0, model);
  874. if ((ret != SUCCESS) || (model.model_data == nullptr)) {
  875. GELOGE(ret, "Load data from file failed. ret = %d", ret);
  876. return ret;
  877. }
  878. ret = ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  879. delete[] static_cast<char *>(model.model_data);
  880. model.model_data = nullptr;
  881. return ret;
  882. }
  883. /**
  884. * @ingroup ge
  885. * @brief Get weight memory size from model file
  886. * @param [in] const void *model_data Offline model buffer
  887. size_t model_size Offline model buffer length
  888. * @param [out] size_t &mem_size Execution memory size
  889. size_t &weight_size Weight memory space size
  890. * @return SUCCESS handle successfully / others handle failed
  891. */
  892. Status GeExecutor::GetMemAndWeightSize(const void *model_data, size_t model_size, size_t &mem_size,
  893. size_t &weight_size) {
  894. GELOGI("Get memory and weight size from data begin.");
  895. if (!isInit_) {
  896. GELOGE(GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  897. return GE_EXEC_NOT_INIT;
  898. }
  899. if (model_data == nullptr) {
  900. GELOGE(PARAM_INVALID, "invalid model data!");
  901. return PARAM_INVALID;
  902. }
  903. ModelData model;
  904. model.model_data = const_cast<void *>(model_data);
  905. model.model_len = static_cast<uint32_t>(model_size);
  906. return ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  907. }
  908. Status GeExecutor::LoadSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  909. SingleOp **single_op) {
  910. return SingleOpManager::GetInstance().GetOpFromModel(model_name, modelData, stream, single_op);
  911. }
  912. Status GeExecutor::LoadDynamicSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  913. DynamicSingleOp **single_op) {
  914. return SingleOpManager::GetInstance().GetDynamicOpFromModel(model_name, modelData, stream, single_op);
  915. }
  916. Status GeExecutor::ExecuteAsync(SingleOp *executor, const std::vector<DataBuffer> &inputs,
  917. std::vector<DataBuffer> &outputs) {
  918. if (executor == nullptr) {
  919. GELOGE(PARAM_INVALID, "param is NULL");
  920. return PARAM_INVALID;
  921. }
  922. return executor->ExecuteAsync(inputs, outputs);
  923. }
  924. ge::Status GeExecutor::ExecuteAsync(DynamicSingleOp *executor, const vector<GeTensorDesc> &input_desc,
  925. const vector<DataBuffer> &inputs, vector<GeTensorDesc> &output_desc,
  926. vector<DataBuffer> &outputs) {
  927. GE_CHECK_NOTNULL(executor);
  928. return executor->ExecuteAsync(input_desc, inputs, output_desc, outputs);
  929. }
  930. Status GeExecutor::ReleaseSingleOpResource(void *stream) {
  931. return SingleOpManager::GetInstance().ReleaseResource(stream);
  932. }
  933. Status GeExecutor::GetBatchInfoSize(uint32_t model_id, size_t &shape_count) {
  934. std::vector<std::vector<int64_t>> batch_info;
  935. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  936. Status ret = GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  937. if (ret != SUCCESS) {
  938. GELOGE(ret, "Calc batch info size failed. ret = %d", ret);
  939. return ret;
  940. }
  941. if (batch_info.empty()) {
  942. shape_count = kStaticBatchInfoSize;
  943. } else {
  944. shape_count = batch_info.size();
  945. }
  946. return SUCCESS;
  947. }
  948. Status GeExecutor::GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info) {
  949. GELOGI("Begin to GetOrigInputInfo.");
  950. if (!isInit_) {
  951. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  952. return GE_EXEC_NOT_INIT;
  953. }
  954. Status ret = GraphExecutor::GetOrigInputInfo(model_id, index, orig_input_info);
  955. if (ret != SUCCESS) {
  956. GELOGE(ret, "GetOrigInputInfo failed.");
  957. return ret;
  958. }
  959. GELOGI("GetOrigInputInfo succ.");
  960. return SUCCESS;
  961. }
  962. Status GeExecutor::GetAllAippInputOutputDims(uint32_t model_id, uint32_t index,
  963. std::vector<InputOutputDims> &input_dims,
  964. std::vector<InputOutputDims> &output_dims) {
  965. GELOGI("Begin to GetAllAippInputOutputDims.");
  966. if (!isInit_) {
  967. GELOGE(GE_EXEC_NOT_INIT, "not inited yet!");
  968. return GE_EXEC_NOT_INIT;
  969. }
  970. Status ret = GraphExecutor::GetAllAippInputOutputDims(model_id, index, input_dims, output_dims);
  971. if (ret != SUCCESS) {
  972. GELOGE(ret, "GetAllAippInputOutputDims failed.");
  973. return ret;
  974. }
  975. GELOGI("GetAllAippInputOutputDims succ.");
  976. return SUCCESS;
  977. }
  978. Status GeExecutor::GetOpDescInfo(uint32_t device_id, uint32_t stream_id, uint32_t task_id, OpDescInfo &op_desc_info) {
  979. GELOGI("Begin to GetOpDescInfo.");
  980. Status ret = GraphExecutor::GetOpDescInfo(device_id, stream_id, task_id, op_desc_info);
  981. if (ret != SUCCESS) {
  982. GELOGE(ret, "GetOpDescInfo failed.");
  983. return ret;
  984. }
  985. GELOGI("GetOpDescInfo succ.");
  986. return SUCCESS;
  987. }
  988. Status GeExecutor::SetDump(const DumpConfig &dump_config) {
  989. GELOGI("Start to set dump config");
  990. auto ret = DumpManager::GetInstance().SetDumpConf(dump_config);
  991. if (ret != SUCCESS) {
  992. GELOGE(ret, "Set dump conf failed");
  993. return ret;
  994. }
  995. GELOGI("Set dump config succ.");
  996. return SUCCESS;
  997. }
  998. } // namespace ge

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示