You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

ge_executor.cc 44 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180
  1. /**
  2. * Copyright 2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "executor/ge_executor.h"
  17. #include <cce/cce.h>
  18. #include <cce/compiler_stub.h>
  19. #include <ctime>
  20. #include <iostream>
  21. #include "common/debug/log.h"
  22. #include "common/ge/ge_util.h"
  23. #include "common/helper/model_helper.h"
  24. #include "common/profiling/profiling_manager.h"
  25. #include "common/dump/dump_manager.h"
  26. #include "common/util.h"
  27. #include "framework/common/debug/ge_log.h"
  28. #include "framework/common/util.h"
  29. #include "graph/execute/graph_execute.h"
  30. #include "graph/load/graph_loader.h"
  31. #include "graph/load/new_model_manager/davinci_model_parser.h"
  32. #include "graph/load/new_model_manager/model_manager.h"
  33. #include "graph/manager/graph_mem_allocator.h"
  34. #include "graph/model.h"
  35. #include "graph/utils/graph_utils.h"
  36. #include "mmpa/mmpa_api.h"
  37. #include "single_op/single_op_manager.h"
  38. #include "graph/manager/graph_var_manager.h"
  39. #include "graph/load/new_model_manager/davinci_model.h"
  40. #include "opskernel_manager/ops_kernel_builder_manager.h"
  41. #include "graph/opsproto_manager.h"
  42. #include "ge_local_engine/engine/host_cpu_engine.h"
  43. using std::string;
  44. using std::vector;
  45. namespace {
  46. const size_t kDynamicBatchSizeVecSize = 1;
  47. const size_t kStaticBatchInfoSize = 1;
  48. const size_t kDynamicImageSizeVecSize = 2;
  49. const size_t kDynamicImageSizeInputSize = 2;
  50. const char *const kBatchLabel = "Batch_";
  51. void GetGeTensorDescFromDomiInfo(std::vector<ge::TensorDesc> &ge_descs,
  52. const std::vector<ge::InputOutputDescInfo> &domi_descs,
  53. const std::vector<uint32_t> &formats) {
  54. uint32_t idx = 0;
  55. for (auto desc_item : domi_descs) {
  56. ge::TensorDesc ge_desc;
  57. ge_desc.SetName(desc_item.name);
  58. ge_desc.SetDataType(static_cast<ge::DataType>(desc_item.data_type));
  59. ge_desc.SetFormat(static_cast<ge::Format>(formats[idx]));
  60. std::vector<int64_t> shape_dims;
  61. for (auto dim : desc_item.shape_info.dims) {
  62. shape_dims.push_back(dim);
  63. }
  64. ge::Shape ge_shape(shape_dims);
  65. ge_desc.SetShape(ge_shape);
  66. ge_desc.SetSize(desc_item.size);
  67. ge_desc.SetShapeRange(desc_item.shape_info.shape_ranges);
  68. ge_descs.emplace_back(ge_desc);
  69. ++idx;
  70. }
  71. }
  72. void GetDomiInputData(const ge::RunModelData &input_data, ge::InputData &inputs) {
  73. inputs.index = input_data.index;
  74. inputs.model_id = input_data.modelId;
  75. inputs.timestamp = input_data.timestamp;
  76. inputs.timeout = input_data.timeout;
  77. inputs.request_id = input_data.request_id;
  78. for (const auto &data_item : input_data.blobs) {
  79. ge::DataBuffer dataBuf{data_item.data, data_item.length, data_item.isDataSupportMemShare};
  80. inputs.blobs.emplace_back(dataBuf);
  81. }
  82. }
  83. void GetDomiOutputData(const ge::RunModelData &output_data, ge::OutputData &outputs) {
  84. outputs.index = output_data.index;
  85. outputs.model_id = output_data.modelId;
  86. for (const auto &data_item : output_data.blobs) {
  87. ge::DataBuffer dataBuf(data_item.data, data_item.length, data_item.isDataSupportMemShare);
  88. outputs.blobs.emplace_back(dataBuf);
  89. }
  90. }
  91. void SetDynamicInputDataFlag(const ge::RunModelData &input_data, const std::vector<std::vector<int64_t>> batch_info,
  92. ge::InputData &inputs) {
  93. inputs.is_dynamic_batch = true;
  94. std::string batch_label;
  95. size_t match_idx = 0;
  96. for (size_t i = 0; i < batch_info.size(); ++i) {
  97. // dynamic_dims
  98. if (input_data.dynamic_dims.size() != 0) {
  99. bool is_match = true;
  100. for (size_t j = 0; j < static_cast<size_t>(input_data.dynamic_dims.size()); ++j) {
  101. if (static_cast<uint64_t>(batch_info[i][j]) != input_data.dynamic_dims[j]) {
  102. is_match = false;
  103. break;
  104. }
  105. }
  106. if (is_match) {
  107. match_idx = i;
  108. break;
  109. }
  110. // dynamic_batch_size
  111. } else if (batch_info[i].size() == kDynamicBatchSizeVecSize &&
  112. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_batch_size)) {
  113. match_idx = i;
  114. break;
  115. // dynamic_image_size
  116. } else if (batch_info[i].size() == kDynamicImageSizeVecSize &&
  117. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_image_height) &&
  118. batch_info[i][1] == static_cast<int64_t>(input_data.dynamic_image_width)) {
  119. match_idx = i;
  120. break;
  121. }
  122. }
  123. batch_label = kBatchLabel + std::to_string(match_idx);
  124. inputs.batch_label = batch_label;
  125. GELOGI("current batch label:%s", batch_label.c_str());
  126. }
  127. bool IsDynamicBatchSizeMatchModel(uint64_t batch_size, const vector<std::vector<int64_t>> &batch_info) {
  128. if (batch_info.empty()) {
  129. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  130. return false;
  131. }
  132. for (auto batch : batch_info) {
  133. if (batch.size() != kDynamicBatchSizeVecSize) {
  134. GELOGE(ge::FAILED, "Dynamic batch param num is %zu, current batch size is %zu.", kDynamicBatchSizeVecSize,
  135. batch.size());
  136. return false;
  137. }
  138. if (batch[0] == static_cast<int64_t>(batch_size)) {
  139. return true;
  140. }
  141. }
  142. GELOGE(ge::FAILED, "Dynamic batch %lu can not match the gear of model.", batch_size);
  143. return false;
  144. }
  145. bool IsDynamicImageSizeMatchModel(uint64_t image_height, uint64_t image_width,
  146. const vector<std::vector<int64_t>> &batch_info) {
  147. if (batch_info.empty()) {
  148. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  149. return false;
  150. }
  151. for (auto resolution : batch_info) {
  152. if (resolution.size() != kDynamicImageSizeVecSize) {
  153. GELOGE(ge::FAILED, "Dynamic resolution param num is %zu, current resolution size is %zu.",
  154. kDynamicImageSizeVecSize, resolution.size());
  155. return false;
  156. }
  157. if (resolution[0] == static_cast<int64_t>(image_height) && resolution[1] == static_cast<int64_t>(image_width)) {
  158. return true;
  159. }
  160. }
  161. GELOGE(ge::FAILED, "Dynamic resolution (%lu,%lu) can not match the gear of model.", image_height, image_width);
  162. return false;
  163. }
  164. bool IsDynmaicDimsSizeMatchModel(const vector<uint64_t> cur_dynamic_dims,
  165. const vector<vector<int64_t>> &batch_info) {
  166. if (batch_info.empty()) {
  167. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  168. return false;
  169. }
  170. bool find_match = false;
  171. for (auto resolution : batch_info) {
  172. if (cur_dynamic_dims.size() != resolution.size()) {
  173. GELOGE(ge::FAILED, "Cur dynamic dims param num is %zu, current resolution size is %zu.",
  174. cur_dynamic_dims.size(), resolution.size());
  175. return false;
  176. }
  177. bool flag = true;
  178. for (std::size_t i = 0; i < resolution.size(); ++i) {
  179. if (cur_dynamic_dims[i] != static_cast<uint64_t>(resolution[i])) {
  180. flag = false;
  181. break;
  182. }
  183. }
  184. if (flag) {
  185. find_match = true;
  186. break;
  187. }
  188. }
  189. if (!find_match) {
  190. GELOGE(ge::FAILED, "choose dynamic dims can not match the gear of model.");
  191. }
  192. return find_match;
  193. }
  194. } // namespace
  195. namespace ge {
  196. bool GeExecutor::isInit_ = false;
  197. class ModelListenerAdapter : public ModelListener {
  198. public:
  199. domi::Status OnComputeDone(uint32_t model_id, uint32_t dataIndex, uint32_t resultCode,
  200. std::vector<ge::OutputTensorInfo> &outputs) {
  201. if (listener == nullptr) {
  202. GELOGE(ge::FAILED, "listener is null.");
  203. return FAILED;
  204. }
  205. return listener->OnComputeDone(model_id, dataIndex, resultCode, outputs);
  206. }
  207. std::shared_ptr<ge::ModelListener> listener;
  208. };
  209. GeExecutor::GeExecutor() {}
  210. Status GeExecutor::Initialize() {
  211. GELOGI("Init GeExecutor begin.");
  212. if (isInit_) {
  213. GELOGW("Already initialized, no need to be initialized again.");
  214. return ge::SUCCESS;
  215. }
  216. std::vector<rtMemType_t> mem_type(1, RT_MEMORY_HBM);
  217. mem_type.push_back(RT_MEMORY_P2P_DDR);
  218. auto ret = MemManager::Instance().Initialize(mem_type);
  219. if (ret != SUCCESS) {
  220. GELOGE(ret, "Memory Manager init failed.");
  221. return ret;
  222. }
  223. GE_CHK_STATUS_RET(OpsKernelBuilderManager::Instance().Initialize({}, false),
  224. "Failed to initialize OpsKernelBuilders");
  225. // Start profiling
  226. Options profiling_options;
  227. profiling_options.device_id = 0;
  228. // job id need to be set, the value is meaningless;
  229. profiling_options.job_id = "1";
  230. ProfilingManager::Instance().Init(profiling_options);
  231. isInit_ = true;
  232. GELOGI("Init GeExecutor over.");
  233. return ge::SUCCESS;
  234. }
  235. Status GeExecutor::Finalize() {
  236. GELOGI("Uninit GeExecutor begin.");
  237. if (isInit_ == false) {
  238. GELOGW("GeExecutor has not been initialized.");
  239. return ge::SUCCESS;
  240. }
  241. (void) OpsKernelBuilderManager::Instance().Finalize();
  242. // Stop profiling
  243. if (ProfilingManager::Instance().ProfilingOn()) {
  244. ProfilingManager::Instance().StopProfiling();
  245. ProfilingManager::Instance().PluginUnInit();
  246. }
  247. GELOGI("Uninit GeExecutor over.");
  248. return ge::SUCCESS;
  249. }
  250. Status GeExecutor::SetDynamicBatchSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  251. uint64_t batch_size) {
  252. if (dynamic_input_addr == nullptr) {
  253. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  254. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  255. }
  256. uint64_t size = sizeof(uint32_t);
  257. if (length < size) {
  258. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID, "Dynamic input size [%lu] is less than [%lu]!", length, size);
  259. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  260. }
  261. if (length >= sizeof(uint64_t)) {
  262. size = sizeof(uint64_t);
  263. }
  264. // Verify whether the input dynamic batch matches the model gear
  265. std::vector<std::vector<int64_t>> batch_info;
  266. std::vector<uint64_t> batch_num{batch_size};
  267. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  268. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  269. if (ret != SUCCESS) {
  270. GELOGE(ret, "Get dynamic input info failed.");
  271. return ret;
  272. }
  273. if (!IsDynamicBatchSizeMatchModel(batch_size, batch_info)) {
  274. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  275. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  276. }
  277. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_BATCH));
  278. if (ret != SUCCESS) {
  279. GELOGE(ret, "Set dynamic size failed");
  280. return ret;
  281. }
  282. // memcpy dynamic_batch_size from host to device
  283. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &batch_size, size, RT_MEMCPY_HOST_TO_DEVICE);
  284. if (rt_ret != RT_ERROR_NONE) {
  285. GELOGE(rt_ret, "memcpy dynamic batch input data failed! ret: 0x%X", rt_ret);
  286. return RT_ERROR_TO_GE_STATUS(rt_ret);
  287. }
  288. return SUCCESS;
  289. }
  290. Status GeExecutor::SetDynamicImageSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  291. uint64_t image_height, uint64_t image_width) {
  292. if (dynamic_input_addr == nullptr) {
  293. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  294. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  295. }
  296. uint64_t dynamic_input_size = kDynamicImageSizeInputSize * sizeof(uint32_t);
  297. if (length < dynamic_input_size) {
  298. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  299. "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  300. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  301. }
  302. uint64_t size = sizeof(uint32_t);
  303. if (length >= kDynamicImageSizeInputSize * sizeof(uint64_t)) {
  304. size = sizeof(uint64_t);
  305. }
  306. // Verify whether the input dynamic resolution matches the model gear
  307. std::vector<std::vector<int64_t>> batch_info;
  308. std::vector<uint64_t> batch_num{image_height, image_width};
  309. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  310. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  311. if (ret != SUCCESS) {
  312. GELOGE(ret, "Get dynamic input info failed.");
  313. return ret;
  314. }
  315. if (!IsDynamicImageSizeMatchModel(image_height, image_width, batch_info)) {
  316. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  317. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  318. }
  319. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_IMAGE));
  320. if (ret != SUCCESS) {
  321. GELOGE(ret, "Set dynamic size failed");
  322. return ret;
  323. }
  324. // Memcpy dynamic resolution height from host to device
  325. rtError_t rt_ret =
  326. rtMemcpy(dynamic_input_addr, size, &image_height, size, RT_MEMCPY_HOST_TO_DEVICE);
  327. if (rt_ret != RT_ERROR_NONE) {
  328. GELOGE(rt_ret, "memcpy dynamic resolution input data failed! ret: 0x%X", rt_ret);
  329. return RT_ERROR_TO_GE_STATUS(rt_ret);
  330. }
  331. uint64_t remain_size = length - size;
  332. // Memcpy dynamic resolution width from host to device
  333. rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size),
  334. remain_size, &image_width, size, RT_MEMCPY_HOST_TO_DEVICE);
  335. if (rt_ret != RT_ERROR_NONE) {
  336. GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
  337. return RT_ERROR_TO_GE_STATUS(rt_ret);
  338. }
  339. return SUCCESS;
  340. }
  341. Status GeExecutor::SetDynamicDims(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  342. const vector<uint64_t> &dynamic_dims) {
  343. if (dynamic_input_addr == nullptr) {
  344. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  345. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  346. }
  347. vector<uint64_t> cur_dynamic_dims;
  348. Status ret = GetCurDynamicDims(model_id, dynamic_dims, cur_dynamic_dims);
  349. if (ret != SUCCESS) {
  350. GELOGE(ret, "Set cur gear dynamic dims failed");
  351. return ret;
  352. }
  353. std::vector<std::vector<int64_t>> batch_info;
  354. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  355. ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  356. if (ret != SUCCESS) {
  357. GELOGE(ret, "Get dynamic input info failed.");
  358. return ret;
  359. }
  360. if (!IsDynmaicDimsSizeMatchModel(cur_dynamic_dims, batch_info)) {
  361. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  362. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  363. }
  364. ret = GraphExecutor::SetDynamicSize(model_id, cur_dynamic_dims, static_cast<int32_t>(DYNAMIC_DIMS));
  365. if (ret != SUCCESS) {
  366. GELOGE(ret, "Set dynamic size failed");
  367. return ret;
  368. }
  369. size_t dynamic_dim_num = cur_dynamic_dims.size();
  370. uint64_t dynamic_input_size = static_cast<uint64_t>(dynamic_dim_num * sizeof(uint32_t));
  371. if (length < dynamic_input_size) {
  372. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  373. "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  374. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  375. }
  376. uint64_t size = sizeof(uint32_t);
  377. if (length >= dynamic_dim_num * sizeof(uint64_t)) {
  378. size = sizeof(uint64_t);
  379. }
  380. rtError_t rt_ret;
  381. for (uint32_t i = 0; i < dynamic_dim_num; ++i) {
  382. // Memcpy dynamic dim[i] from host to device
  383. rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size * i),
  384. length - size * i, &cur_dynamic_dims[i], size, RT_MEMCPY_HOST_TO_DEVICE);
  385. if (rt_ret != RT_ERROR_NONE) {
  386. GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
  387. return RT_ERROR_TO_GE_STATUS(rt_ret);
  388. }
  389. }
  390. return SUCCESS;
  391. }
  392. Status GeExecutor::GetCurDynamicDims(uint32_t model_id, const vector<uint64_t> &dynamic_dims,
  393. vector<uint64_t> &cur_dynamic_dims) {
  394. cur_dynamic_dims.clear();
  395. vector<ge::TensorDesc> input_desc;
  396. vector<ge::TensorDesc> output_desc;
  397. auto ret = GetModelDescInfo(model_id, input_desc, output_desc);
  398. if (ret != ge::SUCCESS) {
  399. GELOGE(ret, "GetModelDescInfo failed.");
  400. return ret;
  401. }
  402. vector<string> user_designate_shape_order;
  403. vector<int64_t> all_data_dims;
  404. ret = GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  405. if (ret != ge::SUCCESS) {
  406. GELOGE(ret, "GetUserDesignateShapeOrder failed.");
  407. return ret;
  408. }
  409. for (auto &data_name : user_designate_shape_order) {
  410. for (auto &desc : input_desc) {
  411. if (desc.GetName() == data_name) {
  412. for (auto dim : desc.GetShape().GetDims()) {
  413. all_data_dims.push_back(dim);
  414. }
  415. break;
  416. }
  417. }
  418. }
  419. if (dynamic_dims.size() != all_data_dims.size()){
  420. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  421. "Dynamic input size [%lu] is not equal with all data dims size [%lu]!",
  422. dynamic_dims.size(), all_data_dims.size());
  423. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  424. }
  425. for (std::size_t i = 0; i < all_data_dims.size(); ++i) {
  426. if (all_data_dims[i] < 0) {
  427. cur_dynamic_dims.push_back(dynamic_dims[i]);
  428. } else if (static_cast<uint64_t>(all_data_dims[i]) != dynamic_dims[i]) {
  429. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID, "Static dims should be same, index: %zu value: %d should be %d",
  430. i, dynamic_dims[i], all_data_dims[i]);
  431. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  432. }
  433. }
  434. return SUCCESS;
  435. }
  436. Status GeExecutor::GetCurShape(const uint32_t model_id, std::vector<int64_t> &batch_info, int32_t &dynamic_type) {
  437. GELOGI("Begin to get current shape");
  438. if (!isInit_) {
  439. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  440. return ACL_ERROR_GE_EXEC_NOT_INIT;
  441. }
  442. Status ret = GraphExecutor::GetCurShape(model_id, batch_info, dynamic_type);
  443. if (ret != SUCCESS) {
  444. GELOGE(ret, "Get current shape failed");
  445. return ret;
  446. }
  447. return SUCCESS;
  448. }
  449. Status GeExecutor::SetDynamicAippData(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  450. const std::vector<kAippDynamicBatchPara> &aippBatchPara,
  451. const kAippDynamicPara &aippParms) {
  452. GELOGI("Enter to SetDynamicAippData.");
  453. if (dynamic_input_addr == nullptr) {
  454. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic aipp input addr is nullptr!");
  455. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  456. }
  457. if (aippBatchPara.empty()) {
  458. GELOGE(ACL_ERROR_GE_AIPP_BATCH_EMPTY, "aippBatchPara is empty.");
  459. return ACL_ERROR_GE_AIPP_BATCH_EMPTY;
  460. }
  461. uint64_t batch_num = aippBatchPara.size();
  462. uint64_t real_aippParms_size = sizeof(kAippDynamicPara) - sizeof(kAippDynamicBatchPara);
  463. uint64_t struct_len = batch_num * sizeof(kAippDynamicBatchPara) + real_aippParms_size;
  464. GELOGI(
  465. "Get acl input dynamic aipp data, model_id is %u, length is %lu,"
  466. "batch num is %lu, struct_len is %lu",
  467. model_id, length, batch_num, struct_len);
  468. if (struct_len > length) {
  469. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  470. "input dynamic aipp param len [%lu] is larger than aipp_data size [%lu]", struct_len, length);
  471. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  472. }
  473. // Memcpy real kAippDynamicBatchPara from host to device
  474. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &aippParms, real_aippParms_size, RT_MEMCPY_HOST_TO_DEVICE);
  475. if (rt_ret != RT_ERROR_NONE) {
  476. GELOGE(rt_ret, "memcpy real_aippParms_size failed! ret: 0x%X", rt_ret);
  477. return RT_ERROR_TO_GE_STATUS(rt_ret);
  478. }
  479. uint64_t remain_len = length - real_aippParms_size;
  480. uint8_t *aipp_batch_para_dev = reinterpret_cast<uint8_t *>(dynamic_input_addr) + real_aippParms_size;
  481. for (uint64_t i = 0; i < batch_num; ++i) {
  482. rt_ret = rtMemcpy(reinterpret_cast<void *>(aipp_batch_para_dev + i * sizeof(kAippDynamicBatchPara)),
  483. (remain_len - i * sizeof(kAippDynamicBatchPara)), &(aippBatchPara[i]),
  484. sizeof(kAippDynamicBatchPara), RT_MEMCPY_HOST_TO_DEVICE);
  485. if (rt_ret != RT_ERROR_NONE) {
  486. GELOGE(rt_ret, "memcpy kAippDynamicBatchPara input data failed! ret: 0x%X", rt_ret);
  487. return RT_ERROR_TO_GE_STATUS(rt_ret);
  488. }
  489. }
  490. return SUCCESS;
  491. }
  492. // Load model
  493. Status GeExecutor::LoadModelOffline(uint32_t &model_id, const std::string &path, const std::string &key,
  494. int32_t priority, std::shared_ptr<ge::ModelListener> listener) {
  495. GELOGI("load model offline begin.");
  496. if (!isInit_) {
  497. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  498. return ACL_ERROR_GE_EXEC_NOT_INIT;
  499. }
  500. string filePath = RealPath(path.c_str());
  501. if (filePath.empty()) {
  502. GELOGE(ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID,
  503. "File path is invalid. please check your text file '%s'.", path.c_str());
  504. return ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID;
  505. }
  506. std::shared_ptr<ModelListenerAdapter> listener_adapter = MakeShared<ModelListenerAdapter>();
  507. if (listener_adapter == nullptr) {
  508. GELOGE(ACL_ERROR_GE_MEMORY_ALLOCATION, "ModelListenerAdapter make shared failed!");
  509. return ACL_ERROR_GE_MEMORY_ALLOCATION;
  510. }
  511. listener_adapter->listener = listener;
  512. Status ret = GraphLoader::LoadModelFromFile(path, key, priority, listener_adapter, model_id);
  513. if (ret != SUCCESS) {
  514. GELOGE(ret, "[GeExecutor] LoadModelFromFile failed");
  515. return ACL_ERROR_GE_LOAD_MODEL;
  516. }
  517. return SUCCESS;
  518. }
  519. Status GeExecutor::LoadModel(uint32_t &model_id, const ModelData &model_data,
  520. std::shared_ptr<ge::ModelListener> listener) {
  521. GELOGI("Load model begin.");
  522. if (!isInit_) {
  523. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  524. return ACL_ERROR_GE_EXEC_NOT_INIT;
  525. }
  526. std::shared_ptr<ModelListenerAdapter> listener_adapter = MakeShared<ModelListenerAdapter>();
  527. if (listener_adapter == nullptr) {
  528. GELOGE(ACL_ERROR_GE_MEMORY_ALLOCATION, "ModelListenerAdapter make shared failed!");
  529. return ACL_ERROR_GE_MEMORY_ALLOCATION;
  530. }
  531. listener_adapter->listener = listener;
  532. Status ret = GraphLoader::LoadModel(model_data, listener_adapter, model_id);
  533. if (ret != SUCCESS) {
  534. GELOGE(ret, "[GeExecutor] LoadModel failed.");
  535. return ACL_ERROR_GE_LOAD_MODEL;
  536. }
  537. return ret;
  538. }
  539. Status GeExecutor::UnloadModel(uint32_t model_id) {
  540. GELOGD("unload model %u begin.", model_id);
  541. if (!isInit_) {
  542. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  543. return ACL_ERROR_GE_EXEC_NOT_INIT;
  544. }
  545. Status ret = GraphLoader::DestroyAicpuSessionForInfer(model_id);
  546. if (ret != SUCCESS) {
  547. GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
  548. return ret;
  549. }
  550. std::shared_ptr<hybrid::HybridDavinciModel> hybrid_davinci_model =
  551. ModelManager::GetInstance()->GetHybridModel(model_id);
  552. if (hybrid_davinci_model != nullptr) {
  553. uint64_t session_id = hybrid_davinci_model->GetSessionId();
  554. VarManagerPool::Instance().RemoveVarManager(session_id);
  555. } else {
  556. std::shared_ptr<DavinciModel> davinci_model = ModelManager::GetInstance()->GetModel(model_id);
  557. if (davinci_model != nullptr) {
  558. uint64_t session_id = davinci_model->GetSessionId();
  559. VarManagerPool::Instance().RemoveVarManager(session_id);
  560. }
  561. }
  562. ret = GraphLoader::UnloadModel(model_id);
  563. if (ret != SUCCESS) {
  564. GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
  565. return ret;
  566. }
  567. return SUCCESS;
  568. }
  569. Status GeExecutor::RunModel(const ge::RunModelData &input_data, ge::RunModelData &output_data) {
  570. GELOGI("run model begin.");
  571. if (!isInit_) {
  572. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  573. return ACL_ERROR_GE_EXEC_NOT_INIT;
  574. }
  575. InputData inputs;
  576. GetDomiInputData(input_data, inputs);
  577. OutputData outputs;
  578. GetDomiOutputData(output_data, outputs);
  579. return GraphExecutor::DataInput(inputs, outputs);
  580. }
  581. // Get input and output descriptor
  582. Status GeExecutor::GetModelDescInfo(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
  583. std::vector<ge::TensorDesc> &output_desc, bool new_model_desc) {
  584. if (!isInit_) {
  585. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  586. return ACL_ERROR_GE_EXEC_NOT_INIT;
  587. }
  588. std::vector<InputOutputDescInfo> input_desc_infos;
  589. std::vector<InputOutputDescInfo> output_desc_infos;
  590. std::vector<uint32_t> input_formats;
  591. std::vector<uint32_t> output_formats;
  592. Status ret = GraphExecutor::GetInputOutputDescInfo(model_id, input_desc_infos, output_desc_infos, input_formats,
  593. output_formats, new_model_desc);
  594. if (ret != domi::SUCCESS) {
  595. GELOGE(ret, "GetInputOutputDescInfo failed. ret = %u", ret);
  596. return ACL_ERROR_GE_GET_TENSOR_INFO;
  597. }
  598. if (input_formats.size() != input_desc_infos.size()) {
  599. GELOGE(ACL_ERROR_GE_PARAM_INVALID,
  600. "input_formats size %zu is not equal to input_desc_infos size %zu.",
  601. input_formats.size(), input_desc_infos.size());
  602. return ACL_ERROR_GE_PARAM_INVALID;
  603. }
  604. if (output_formats.size() != output_desc_infos.size()) {
  605. GELOGE(ACL_ERROR_GE_PARAM_INVALID, "output_formats size %zu is not equal to output_desc_infos size %zu.",
  606. output_formats.size(), output_desc_infos.size());
  607. return ACL_ERROR_GE_PARAM_INVALID;
  608. }
  609. // Transfer data to TensorDesc
  610. GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
  611. GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
  612. return ge::SUCCESS;
  613. }
  614. ///
  615. /// @ingroup ge
  616. /// @brief Get dynamic batch_info
  617. /// @param [in] model_id
  618. /// @param [out] batch_info
  619. /// @param [out] dynamic_type
  620. /// @return execute result
  621. ///
  622. Status GeExecutor::GetDynamicBatchInfo(uint32_t model_id, std::vector<std::vector<int64_t>> &batch_info,
  623. int32_t &dynamic_type) {
  624. if (!isInit_) {
  625. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  626. return ACL_ERROR_GE_EXEC_NOT_INIT;
  627. }
  628. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  629. if (ret != SUCCESS) {
  630. GELOGE(ret, "GetDynamicBatchInfo failed.");
  631. return ret;
  632. }
  633. return SUCCESS;
  634. }
  635. ///
  636. /// @ingroup ge
  637. /// @brief Get combined dynamic dims info
  638. /// @param [in] model_id
  639. /// @param [out] batch_info
  640. /// @return execute result
  641. ///
  642. Status GeExecutor::GetCombinedDynamicDims(uint32_t model_id, vector<vector<int64_t>> &batch_info) {
  643. GELOGI("Begin to get combined dynamic dims info.");
  644. if (!isInit_) {
  645. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  646. return ACL_ERROR_GE_EXEC_NOT_INIT;
  647. }
  648. Status ret = GraphExecutor::GetCombinedDynamicDims(model_id, batch_info);
  649. if (ret != SUCCESS) {
  650. GELOGE(ret, "GetCombinedDynamicDims failed.");
  651. return ret;
  652. }
  653. GELOGI("Get combined dynamic dims succ.");
  654. return SUCCESS;
  655. }
  656. ///
  657. /// @ingroup ge
  658. /// @brief Get user designeate shape order
  659. /// @param [in] model_id
  660. /// @param [out] user_designate_shape_order
  661. /// @return execute result
  662. ///
  663. Status GeExecutor::GetUserDesignateShapeOrder(uint32_t model_id, vector<string> &user_designate_shape_order) {
  664. if (!isInit_) {
  665. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  666. return ACL_ERROR_GE_EXEC_NOT_INIT;
  667. }
  668. Status ret = GraphExecutor::GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  669. if (ret != SUCCESS) {
  670. GELOGE(ret, "GetUserDesignateShapeOrder failed.");
  671. return ret;
  672. }
  673. return SUCCESS;
  674. }
  675. ///
  676. /// @ingroup ge
  677. /// @brief Get AIPP input format
  678. /// @param [in] model_id
  679. /// @param [in] index
  680. /// @param [out] input_format
  681. /// @return execute result
  682. ///
  683. Status GeExecutor::GetAIPPInfo(uint32_t model_id, uint32_t index, AippConfigInfo &aipp_info) {
  684. GELOGI("Begin to GetAIPPInfo.");
  685. if (!isInit_) {
  686. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  687. return ACL_ERROR_GE_EXEC_NOT_INIT;
  688. }
  689. Status ret = GraphExecutor::GetAIPPInfo(model_id, index, aipp_info);
  690. if (ret != SUCCESS) {
  691. GELOGW("GetAIPPInfo is not success.");
  692. return ret;
  693. }
  694. GELOGI("GetAIPPInfo succ.");
  695. return SUCCESS;
  696. }
  697. Status GeExecutor::GetAippType(uint32_t model_id, uint32_t index, InputAippType &type, size_t &aipp_index) {
  698. GELOGI("Begin to get aipp type.");
  699. if (!isInit_) {
  700. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  701. return ACL_ERROR_GE_EXEC_NOT_INIT;
  702. }
  703. Status ret = GraphExecutor::GetAippType(model_id, index, type, aipp_index);
  704. if (ret != SUCCESS) {
  705. GELOGW("Get aipp type is not success.");
  706. return ret;
  707. }
  708. GELOGI("Get aipp type success.");
  709. return SUCCESS;
  710. }
  711. Status GeExecutor::GetModelAttr(uint32_t model_id, std::vector<std::string> &dynamic_output_shape_info) {
  712. if (!isInit_) {
  713. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  714. return ACL_ERROR_GE_EXEC_NOT_INIT;
  715. }
  716. Status ret = GraphExecutor::GetModelAttr(model_id, dynamic_output_shape_info);
  717. if (ret != SUCCESS) {
  718. GELOGE(ret, "Get dynamic batch output shape info failed.");
  719. return ret;
  720. }
  721. return SUCCESS;
  722. }
  723. Status GeExecutor::GetModelDescInfoForZeroCopy(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
  724. std::vector<TensorDesc> &output_desc) {
  725. GELOGI("get model desc info for zero copy begin.");
  726. if (!isInit_) {
  727. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  728. return ACL_ERROR_GE_EXEC_NOT_INIT;
  729. }
  730. std::vector<InputOutputDescInfo> input_desc_infos;
  731. std::vector<InputOutputDescInfo> output_desc_infos;
  732. std::vector<uint32_t> input_formats;
  733. std::vector<uint32_t> output_formats;
  734. Status ret = GraphExecutor::GetInputOutputDescInfoForZeroCopy(model_id, input_desc_infos, output_desc_infos,
  735. input_formats, output_formats);
  736. if (ret != domi::SUCCESS) {
  737. GELOGE(ret, "Get DescInfo from zero copy failed. ret = %u", ret);
  738. return ACL_ERROR_GE_GET_TENSOR_INFO;
  739. }
  740. if (input_formats.size() != input_desc_infos.size()) {
  741. GELOGE(ACL_ERROR_GE_PARAM_INVALID, "input_formats.size() != input_desc_infos.size().");
  742. return ACL_ERROR_GE_PARAM_INVALID;
  743. }
  744. if (output_formats.size() != output_desc_infos.size()) {
  745. GELOGE(ACL_ERROR_GE_PARAM_INVALID, "output_formats.size() != output_desc_infos.size().");
  746. return ACL_ERROR_GE_PARAM_INVALID;
  747. }
  748. GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
  749. GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
  750. GELOGI("get model desc info from zero copy end.");
  751. return ge::SUCCESS;
  752. }
  753. Status GeExecutor::CommandHandle(const Command &command) {
  754. Status ret = GraphLoader::CommandHandle(command);
  755. if (ret != SUCCESS) {
  756. GELOGE(ACL_ERROR_GE_COMMAND_HANDLE, "CommandHandle: Command Handle failed.");
  757. return ACL_ERROR_GE_COMMAND_HANDLE;
  758. }
  759. return SUCCESS;
  760. }
  761. Status GeExecutor::GetMaxUsedMemory(uint32_t model_id, uint32_t &max_size) {
  762. GELOGI("Get max used memory begin.");
  763. if (!isInit_) {
  764. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  765. return ACL_ERROR_GE_EXEC_NOT_INIT;
  766. }
  767. uint64_t max_mem_size = 0;
  768. Status ret = GraphLoader::GetMaxUsedMemory(model_id, max_mem_size);
  769. max_size = static_cast<uint32_t>(max_mem_size);
  770. return ret;
  771. }
  772. /**
  773. * @ingroup ge
  774. * @brief Load data from model file to memory
  775. * @param [in] const std::string &path: Offline model file path
  776. * @param [out] domi::ModelData &model_data: Offline model memory data
  777. * @return SUCCESS handle successfully / others handle failed
  778. */
  779. Status GeExecutor::LoadDataFromFile(const std::string &path, ModelData &model_data) {
  780. GELOGI("Load data from file begin.");
  781. if (!isInit_) {
  782. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  783. return ACL_ERROR_GE_EXEC_NOT_INIT;
  784. }
  785. string filePath = RealPath(path.c_str());
  786. if (filePath.empty()) {
  787. GELOGE(ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID,
  788. "File path is invalid. please check your text file '%s'.", path.c_str());
  789. return ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID;
  790. }
  791. GELOGI("load modelData from file: %s.", path.c_str());
  792. std::string key_path;
  793. int32_t priority = 0;
  794. Status ret = GraphLoader::LoadDataFromFile(path, key_path, priority, model_data);
  795. if (ret != SUCCESS) {
  796. if (model_data.model_data != nullptr) {
  797. delete[] static_cast<char *>(model_data.model_data);
  798. model_data.model_data = nullptr;
  799. }
  800. }
  801. return ret;
  802. }
  803. /**
  804. * @ingroup ge
  805. * @brief Load model from offline model memory data
  806. * @param [in] domi::ModelData &model_data: Offline model data
  807. void *dev_ptr: Input/Output memory start address
  808. size_t memsize: Input/Output memory length
  809. void *weight_ptr: Weight memory start address
  810. size_t weightsize: Weight memory length
  811. * @param [out] uint32_t &model_id: identification after model loading
  812. * @return SUCCESS handle successfully / others handle failed
  813. */
  814. Status GeExecutor::LoadModelFromData(uint32_t &model_id, const ModelData &model_data, void *dev_ptr, size_t mem_size,
  815. void *weight_ptr, size_t weight_size) {
  816. if (!isInit_) {
  817. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  818. return ACL_ERROR_GE_EXEC_NOT_INIT;
  819. }
  820. return GraphLoader::LoadModelFromData(model_id, model_data, dev_ptr, mem_size, weight_ptr, weight_size);
  821. }
  822. /**
  823. * @ingroup ge
  824. * @brief Load task list from ModelData with queue.
  825. * @param [out] model_id: model id allocate from manager.
  826. * @param [in] ge_model_data: Model data load from offline model.
  827. * @param [in] input_queue_ids: input queue ids create from user.
  828. * @param [in] output_queue_ids: input queue ids create from user.
  829. * @return: 0 for success / others for fail
  830. */
  831. Status GeExecutor::LoadModelWithQ(uint32_t &model_id, const ModelData &model_data,
  832. const std::vector<uint32_t> &input_queue_ids,
  833. const std::vector<uint32_t> &output_queue_ids) {
  834. GELOGI("Load model with queue begin.");
  835. if (!isInit_) {
  836. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  837. return ACL_ERROR_GE_EXEC_NOT_INIT;
  838. }
  839. return GraphLoader::LoadModelWithQ(model_id, model_data, input_queue_ids, output_queue_ids);
  840. }
  841. /**
  842. * @ingroup ge
  843. * @brief Synchronous execution of offline model(Do not create thread)
  844. * @param [in] uint32_t model_id: Model ID to execute
  845. void* stream: stream to execute
  846. const domi::InputData *input_data: Model input data
  847. bool async_mode: is asynchronize mode.
  848. * @param [out] domi::OutputData *output_data: Model output data
  849. * @return SUCCESS handle successfully / others handle failed
  850. */
  851. Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
  852. ge::RunModelData &run_output_data, bool async_mode) {
  853. std::vector<GeTensorDesc> input_desc = {};
  854. std::vector<GeTensorDesc> output_desc = {};
  855. return ExecModel(model_id, stream, run_input_data, input_desc, run_output_data, output_desc, async_mode);
  856. }
  857. /**
  858. * @ingroup ge
  859. * @brief Synchronous execution of offline model(Do not create thread)
  860. * @param [in] uint32_t model_id: Model ID to execute
  861. void* stream: stream to execute
  862. const domi::InputData *input_data: Model input data
  863. const std::vector<GeTensorDesc> &input_desc: Description of model input data
  864. bool async_mode: is asynchronize mode
  865. * @param [out] domi::OutputData *output_data: Model output data
  866. * @param [out] std::vector<GeTensorDesc> &output_desc: Description of model output data
  867. * @return SUCCESS handle successfully / others handle failed
  868. */
  869. Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
  870. const std::vector<GeTensorDesc> &input_desc, ge::RunModelData &run_output_data,
  871. std::vector<GeTensorDesc> &output_desc, bool async_mode) {
  872. if (!isInit_) {
  873. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  874. return ACL_ERROR_GE_EXEC_NOT_INIT;
  875. }
  876. InputData input_data;
  877. OutputData output_data;
  878. GetDomiInputData(run_input_data, input_data);
  879. GetDomiOutputData(run_output_data, output_data);
  880. if ((run_input_data.dynamic_batch_size != 0) || (run_input_data.dynamic_image_width != 0) ||
  881. (run_input_data.dynamic_image_height != 0) || (run_input_data.dynamic_dims.size() != 0)) {
  882. std::vector<std::vector<int64_t>> batch_info;
  883. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  884. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  885. if (ret != SUCCESS) {
  886. GELOGE(ret, "Get dynamic input info failed.");
  887. return ret;
  888. }
  889. if (!batch_info.empty()) {
  890. SetDynamicInputDataFlag(run_input_data, batch_info, input_data);
  891. }
  892. }
  893. return GraphLoader::ExecuteModel(model_id, stream, async_mode, input_data, input_desc, output_data, output_desc);
  894. }
  895. /**
  896. * @ingroup ge
  897. * @brief Get weight memory size from model file
  898. * @param [in] const std::string &path: Offline model file path
  899. * @param [out] size_t &mem_size Execution memory size
  900. size_t &weight_size Weight memory space size
  901. * @return SUCCESS handle successfully / others handle failed
  902. */
  903. Status GeExecutor::GetMemAndWeightSize(const std::string &path, size_t &mem_size, size_t &weight_size) {
  904. GELOGI("Get memory and weight size from file begin.");
  905. if (!isInit_) {
  906. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  907. return ACL_ERROR_GE_EXEC_NOT_INIT;
  908. }
  909. ModelData model;
  910. std::string key;
  911. Status ret = ge::GraphLoader::LoadDataFromFile(path, key, 0, model);
  912. if ((ret != SUCCESS) || (model.model_data == nullptr)) {
  913. GELOGE(ret, "Load data from file failed. ret = %d", ret);
  914. return ret;
  915. }
  916. ret = ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  917. delete[] static_cast<char *>(model.model_data);
  918. model.model_data = nullptr;
  919. return ret;
  920. }
  921. /**
  922. * @ingroup ge
  923. * @brief Get weight memory size from model file
  924. * @param [in] const void *model_data Offline model buffer
  925. size_t model_size Offline model buffer length
  926. * @param [out] size_t &mem_size Execution memory size
  927. size_t &weight_size Weight memory space size
  928. * @return SUCCESS handle successfully / others handle failed
  929. */
  930. Status GeExecutor::GetMemAndWeightSize(const void *model_data, size_t model_size, size_t &mem_size,
  931. size_t &weight_size) {
  932. GELOGI("Get memory and weight size from data begin.");
  933. if (!isInit_) {
  934. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  935. return ACL_ERROR_GE_EXEC_NOT_INIT;
  936. }
  937. if (model_data == nullptr) {
  938. GELOGE(ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID, "invalid model data!");
  939. return ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID;
  940. }
  941. ModelData model;
  942. model.model_data = const_cast<void *>(model_data);
  943. model.model_len = static_cast<uint32_t>(model_size);
  944. return ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  945. }
  946. Status GeExecutor::LoadSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  947. SingleOp **single_op) {
  948. return LoadSingleOpV2(model_name, modelData, stream, single_op, 0);
  949. }
  950. Status GeExecutor::LoadSingleOpV2(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  951. SingleOp **single_op, const uint64_t model_id) {
  952. return SingleOpManager::GetInstance().GetOpFromModel(model_name, modelData, stream, single_op, model_id);
  953. }
  954. Status GeExecutor::LoadDynamicSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  955. DynamicSingleOp **single_op) {
  956. return LoadDynamicSingleOpV2(model_name, modelData, stream, single_op, 0);
  957. }
  958. Status GeExecutor::LoadDynamicSingleOpV2(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  959. DynamicSingleOp **single_op, const uint64_t model_id) {
  960. return SingleOpManager::GetInstance().GetDynamicOpFromModel(model_name, modelData, stream, single_op, model_id);
  961. }
  962. Status GeExecutor::ExecuteAsync(SingleOp *executor, const std::vector<DataBuffer> &inputs,
  963. std::vector<DataBuffer> &outputs) {
  964. if (executor == nullptr) {
  965. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "param is NULL");
  966. return ACL_ERROR_GE_EXEC_NOT_INIT;
  967. }
  968. return executor->ExecuteAsync(inputs, outputs);
  969. }
  970. ge::Status GeExecutor::ExecuteAsync(DynamicSingleOp *executor, const vector<GeTensorDesc> &input_desc,
  971. const vector<DataBuffer> &inputs, vector<GeTensorDesc> &output_desc,
  972. vector<DataBuffer> &outputs) {
  973. GE_CHECK_NOTNULL(executor);
  974. return executor->ExecuteAsync(input_desc, inputs, output_desc, outputs);
  975. }
  976. Status GeExecutor::ReleaseSingleOpResource(void *stream) {
  977. ModelManager::GetInstance()->ClearAicpuSo();
  978. return SingleOpManager::GetInstance().ReleaseResource(stream);
  979. }
  980. Status GeExecutor::GetDeviceIdByModelId(uint32_t model_id, uint32_t &device_id) {
  981. auto model_manager = ModelManager::GetInstance();
  982. GE_CHECK_NOTNULL(model_manager);
  983. auto davinci_model = model_manager->GetModel(model_id);
  984. if (davinci_model == nullptr) {
  985. GELOGE(ACL_ERROR_GE_EXEC_MODEL_ID_INVALID, "Model id: %d is invaild or model is not loaded.", model_id);
  986. return ACL_ERROR_GE_EXEC_MODEL_ID_INVALID;
  987. }
  988. device_id = davinci_model->GetDeviceId();
  989. return SUCCESS;
  990. }
  991. Status GeExecutor::GetBatchInfoSize(uint32_t model_id, size_t &shape_count) {
  992. std::vector<std::vector<int64_t>> batch_info;
  993. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  994. Status ret = GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  995. if (ret != SUCCESS) {
  996. GELOGE(ret, "Calc batch info size failed. ret = %d", ret);
  997. return ret;
  998. }
  999. if (batch_info.empty()) {
  1000. shape_count = kStaticBatchInfoSize;
  1001. } else {
  1002. shape_count = batch_info.size();
  1003. }
  1004. return SUCCESS;
  1005. }
  1006. Status GeExecutor::GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info) {
  1007. GELOGI("Begin to GetOrigInputInfo.");
  1008. if (!isInit_) {
  1009. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  1010. return ACL_ERROR_GE_EXEC_NOT_INIT;
  1011. }
  1012. Status ret = GraphExecutor::GetOrigInputInfo(model_id, index, orig_input_info);
  1013. if (ret != SUCCESS) {
  1014. GELOGE(ret, "GetOrigInputInfo failed.");
  1015. return ret;
  1016. }
  1017. GELOGI("GetOrigInputInfo succ.");
  1018. return SUCCESS;
  1019. }
  1020. Status GeExecutor::GetAllAippInputOutputDims(uint32_t model_id, uint32_t index,
  1021. std::vector<InputOutputDims> &input_dims,
  1022. std::vector<InputOutputDims> &output_dims) {
  1023. GELOGI("Begin to GetAllAippInputOutputDims.");
  1024. if (!isInit_) {
  1025. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  1026. return ACL_ERROR_GE_EXEC_NOT_INIT;
  1027. }
  1028. Status ret = GraphExecutor::GetAllAippInputOutputDims(model_id, index, input_dims, output_dims);
  1029. if (ret != SUCCESS) {
  1030. GELOGE(ret, "GetAllAippInputOutputDims failed.");
  1031. return ret;
  1032. }
  1033. GELOGI("GetAllAippInputOutputDims succ.");
  1034. return SUCCESS;
  1035. }
  1036. Status GeExecutor::GetOpDescInfo(uint32_t device_id, uint32_t stream_id, uint32_t task_id, OpDescInfo &op_desc_info) {
  1037. GELOGI("Begin to GetOpDescInfo.");
  1038. Status ret = GraphExecutor::GetOpDescInfo(device_id, stream_id, task_id, op_desc_info);
  1039. if (ret != SUCCESS) {
  1040. GELOGE(ret, "GetOpDescInfo failed.");
  1041. return ret;
  1042. }
  1043. GELOGI("GetOpDescInfo succ.");
  1044. return SUCCESS;
  1045. }
  1046. Status GeExecutor::SetDump(const DumpConfig &dump_config) {
  1047. GELOGI("Start to set dump config");
  1048. auto ret = DumpManager::GetInstance().SetDumpConf(dump_config);
  1049. if (ret != SUCCESS) {
  1050. GELOGE(ret, "Set dump conf failed");
  1051. return ret;
  1052. }
  1053. GELOGI("Set dump config successfully");
  1054. return SUCCESS;
  1055. }
  1056. } // namespace ge

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示