You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

ge_executor.cc 39 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
5 years ago
5 years ago
5 years ago
5 years ago
4 years ago
5 years ago
5 years ago
4 years ago
5 years ago
4 years ago
4 years ago
4 years ago
4 years ago
5 years ago
12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049
  1. /**
  2. * Copyright 2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "executor/ge_executor.h"
  17. #include <cce/cce.h>
  18. #include <cce/compiler_stub.h>
  19. #include <ctime>
  20. #include <iostream>
  21. #include "common/debug/log.h"
  22. #include "common/ge/ge_util.h"
  23. #include "common/helper/model_helper.h"
  24. #include "common/profiling/profiling_manager.h"
  25. #include "common/dump/dump_manager.h"
  26. #include "common/util.h"
  27. #include "framework/common/debug/ge_log.h"
  28. #include "framework/common/util.h"
  29. #include "graph/execute/graph_execute.h"
  30. #include "graph/load/graph_loader.h"
  31. #include "graph/load/model_manager/davinci_model_parser.h"
  32. #include "graph/load/model_manager/model_manager.h"
  33. #include "graph/manager/graph_mem_allocator.h"
  34. #include "graph/model.h"
  35. #include "graph/utils/graph_utils.h"
  36. #include "mmpa/mmpa_api.h"
  37. #include "single_op/single_op_manager.h"
  38. #include "graph/manager/graph_var_manager.h"
  39. #include "graph/load/model_manager/davinci_model.h"
  40. #include "opskernel_manager/ops_kernel_builder_manager.h"
  41. using std::string;
  42. using std::vector;
  43. namespace {
  44. const size_t kDynamicBatchSizeVecSize = 1;
  45. const size_t kStaticBatchInfoSize = 1;
  46. const size_t kDynamicImageSizeVecSize = 2;
  47. const size_t kDynamicImageSizeInputSize = 2;
  48. const char *const kBatchLabel = "Batch_";
  49. void GetGeTensorDescFromDomiInfo(std::vector<ge::TensorDesc> &ge_descs,
  50. const std::vector<ge::InputOutputDescInfo> &domi_descs,
  51. const std::vector<uint32_t> &formats) {
  52. uint32_t idx = 0;
  53. for (auto desc_item : domi_descs) {
  54. ge::TensorDesc ge_desc;
  55. ge_desc.SetName(desc_item.name);
  56. ge_desc.SetDataType(static_cast<ge::DataType>(desc_item.data_type));
  57. ge_desc.SetFormat(static_cast<ge::Format>(formats[idx]));
  58. std::vector<int64_t> shape_dims;
  59. for (auto dim : desc_item.shape_info.dims) {
  60. shape_dims.push_back(dim);
  61. }
  62. ge::Shape ge_shape(shape_dims);
  63. ge_desc.SetShape(ge_shape);
  64. ge_desc.SetSize(desc_item.size);
  65. ge_desc.SetShapeRange(desc_item.shape_info.shape_ranges);
  66. ge_descs.emplace_back(ge_desc);
  67. ++idx;
  68. }
  69. }
  70. void GetDomiInputData(const ge::RunModelData &input_data, ge::InputData &inputs) {
  71. inputs.index = input_data.index;
  72. inputs.model_id = input_data.modelId;
  73. inputs.timestamp = input_data.timestamp;
  74. inputs.timeout = input_data.timeout;
  75. inputs.request_id = input_data.request_id;
  76. for (const auto &data_item : input_data.blobs) {
  77. ge::DataBuffer dataBuf{data_item.data, data_item.length, data_item.isDataSupportMemShare};
  78. inputs.blobs.emplace_back(dataBuf);
  79. }
  80. }
  81. void GetDomiOutputData(const ge::RunModelData &output_data, ge::OutputData &outputs) {
  82. outputs.index = output_data.index;
  83. outputs.model_id = output_data.modelId;
  84. for (const auto &data_item : output_data.blobs) {
  85. ge::DataBuffer dataBuf(data_item.data, data_item.length, data_item.isDataSupportMemShare);
  86. outputs.blobs.emplace_back(dataBuf);
  87. }
  88. }
  89. void SetDynamicInputDataFlag(const ge::RunModelData &input_data, const std::vector<std::vector<int64_t>> batch_info,
  90. ge::InputData &inputs) {
  91. inputs.is_dynamic_batch = true;
  92. std::string batch_label;
  93. size_t match_idx = 0;
  94. for (size_t i = 0; i < batch_info.size(); ++i) {
  95. // dynamic_dims
  96. if (input_data.dynamic_dims.size() != 0) {
  97. bool is_match = true;
  98. for (size_t j = 0; j < static_cast<size_t>(input_data.dynamic_dims.size()); ++j) {
  99. if (static_cast<uint64_t>(batch_info[i][j]) != input_data.dynamic_dims[j]) {
  100. is_match = false;
  101. break;
  102. }
  103. }
  104. if (is_match) {
  105. match_idx = i;
  106. break;
  107. }
  108. // dynamic_batch_size
  109. } else if (batch_info[i].size() == kDynamicBatchSizeVecSize &&
  110. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_batch_size)) {
  111. match_idx = i;
  112. break;
  113. // dynamic_image_size
  114. } else if (batch_info[i].size() == kDynamicImageSizeVecSize &&
  115. batch_info[i][0] == static_cast<int64_t>(input_data.dynamic_image_height) &&
  116. batch_info[i][1] == static_cast<int64_t>(input_data.dynamic_image_width)) {
  117. match_idx = i;
  118. break;
  119. }
  120. }
  121. batch_label = kBatchLabel + std::to_string(match_idx);
  122. inputs.batch_label = batch_label;
  123. GELOGI("current batch label:%s", batch_label.c_str());
  124. }
  125. bool IsDynamicBatchSizeMatchModel(uint64_t batch_size, const vector<std::vector<int64_t>> &batch_info) {
  126. if (batch_info.empty()) {
  127. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  128. return false;
  129. }
  130. for (auto batch : batch_info) {
  131. if (batch.size() != kDynamicBatchSizeVecSize) {
  132. GELOGE(ge::FAILED, "Dynamic batch param num is %zu, current batch size is %zu.", kDynamicBatchSizeVecSize,
  133. batch.size());
  134. return false;
  135. }
  136. if (batch[0] == static_cast<int64_t>(batch_size)) {
  137. return true;
  138. }
  139. }
  140. GELOGE(ge::FAILED, "Dynamic batch %lu can not match the gear of model.", batch_size);
  141. return false;
  142. }
  143. bool IsDynamicImageSizeMatchModel(uint64_t image_height, uint64_t image_width,
  144. const vector<std::vector<int64_t>> &batch_info) {
  145. if (batch_info.empty()) {
  146. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  147. return false;
  148. }
  149. for (auto resolution : batch_info) {
  150. if (resolution.size() != kDynamicImageSizeVecSize) {
  151. GELOGE(ge::FAILED, "Dynamic resolution param num is %zu, current resolution size is %zu.",
  152. kDynamicImageSizeVecSize, resolution.size());
  153. return false;
  154. }
  155. if (resolution[0] == static_cast<int64_t>(image_height) && resolution[1] == static_cast<int64_t>(image_width)) {
  156. return true;
  157. }
  158. }
  159. GELOGE(ge::FAILED, "Dynamic resolution (%lu,%lu) can not match the gear of model.", image_height, image_width);
  160. return false;
  161. }
  162. bool IsDynmaicDimsSizeMatchModel(const vector<uint64_t> cur_dynamic_dims,
  163. const vector<vector<int64_t>> &batch_info) {
  164. if (batch_info.empty()) {
  165. GELOGE(ge::FAILED, "Dynamic batch info is empty.");
  166. return false;
  167. }
  168. bool find_match = false;
  169. for (auto resolution : batch_info) {
  170. if (cur_dynamic_dims.size() != resolution.size()) {
  171. GELOGE(ge::FAILED, "Cur dynamic dims param num is %zu, current resolution size is %zu.",
  172. cur_dynamic_dims.size(), resolution.size());
  173. return false;
  174. }
  175. bool flag = true;
  176. for (std::size_t i = 0; i < resolution.size(); ++i) {
  177. if (cur_dynamic_dims[i] != static_cast<uint64_t>(resolution[i])) {
  178. flag = false;
  179. break;
  180. }
  181. }
  182. if (flag) {
  183. find_match = true;
  184. break;
  185. }
  186. }
  187. if (!find_match) {
  188. GELOGE(ge::FAILED, "choose dynamic dims can not match the gear of model.");
  189. }
  190. return find_match;
  191. }
  192. } // namespace
  193. namespace ge {
  194. bool GeExecutor::isInit_ = false;
  195. GeExecutor::GeExecutor() {}
  196. Status GeExecutor::Initialize() {
  197. GELOGI("Init GeExecutor begin.");
  198. if (isInit_) {
  199. GELOGW("Already initialized, no need to be initialized again.");
  200. return ge::SUCCESS;
  201. }
  202. std::vector<rtMemType_t> mem_type(1, RT_MEMORY_HBM);
  203. mem_type.push_back(RT_MEMORY_P2P_DDR);
  204. auto ret = MemManager::Instance().Initialize(mem_type);
  205. if (ret != SUCCESS) {
  206. GELOGE(ret, "Memory Manager init failed.");
  207. return ret;
  208. }
  209. GE_CHK_STATUS_RET(OpsKernelBuilderManager::Instance().Initialize({}, false),
  210. "Failed to initialize OpsKernelBuilders.");
  211. // Start profiling
  212. Options profiling_options;
  213. profiling_options.device_id = 0;
  214. // job id need to be set, the value is meaningless;
  215. profiling_options.job_id = "1";
  216. ProfilingManager::Instance().Init(profiling_options);
  217. isInit_ = true;
  218. GELOGI("Init GeExecutor over.");
  219. return ge::SUCCESS;
  220. }
  221. Status GeExecutor::Finalize() {
  222. GELOGI("Uninit GeExecutor begin.");
  223. if (isInit_ == false) {
  224. GELOGW("GeExecutor has not been initialized.");
  225. return ge::SUCCESS;
  226. }
  227. (void) OpsKernelBuilderManager::Instance().Finalize();
  228. // Stop profiling
  229. if (ProfilingManager::Instance().ProfilingOn()) {
  230. ProfilingManager::Instance().StopProfiling();
  231. ProfilingManager::Instance().PluginUnInit();
  232. }
  233. GELOGI("Uninit GeExecutor over.");
  234. return ge::SUCCESS;
  235. }
  236. Status GeExecutor::SetDynamicBatchSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  237. uint64_t batch_size) {
  238. if (dynamic_input_addr == nullptr) {
  239. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  240. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  241. }
  242. uint64_t size = sizeof(uint32_t);
  243. if (length < size) {
  244. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID, "Dynamic input size [%lu] is less than [%lu]!", length, size);
  245. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  246. }
  247. if (length >= sizeof(uint64_t)) {
  248. size = sizeof(uint64_t);
  249. }
  250. // Verify whether the input dynamic batch matches the model gear
  251. std::vector<std::vector<int64_t>> batch_info;
  252. std::vector<uint64_t> batch_num{batch_size};
  253. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  254. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  255. if (ret != SUCCESS) {
  256. GELOGE(ret, "Get dynamic input info failed.");
  257. return ret;
  258. }
  259. if (!IsDynamicBatchSizeMatchModel(batch_size, batch_info)) {
  260. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  261. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  262. }
  263. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_BATCH));
  264. if (ret != SUCCESS) {
  265. GELOGE(ret, "Set dynamic size failed");
  266. return ret;
  267. }
  268. // memcpy dynamic_batch_size from host to device
  269. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &batch_size, size, RT_MEMCPY_HOST_TO_DEVICE);
  270. if (rt_ret != RT_ERROR_NONE) {
  271. GELOGE(rt_ret, "memcpy dynamic batch input data failed! ret: 0x%X", rt_ret);
  272. return RT_ERROR_TO_GE_STATUS(rt_ret);
  273. }
  274. return SUCCESS;
  275. }
  276. Status GeExecutor::SetDynamicImageSize(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  277. uint64_t image_height, uint64_t image_width) {
  278. if (dynamic_input_addr == nullptr) {
  279. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  280. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  281. }
  282. uint64_t dynamic_input_size = kDynamicImageSizeInputSize * sizeof(uint32_t);
  283. if (length < dynamic_input_size) {
  284. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  285. "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  286. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  287. }
  288. uint64_t size = sizeof(uint32_t);
  289. if (length >= kDynamicImageSizeInputSize * sizeof(uint64_t)) {
  290. size = sizeof(uint64_t);
  291. }
  292. // Verify whether the input dynamic resolution matches the model gear
  293. std::vector<std::vector<int64_t>> batch_info;
  294. std::vector<uint64_t> batch_num{image_height, image_width};
  295. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  296. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  297. if (ret != SUCCESS) {
  298. GELOGE(ret, "Get dynamic input info failed.");
  299. return ret;
  300. }
  301. if (!IsDynamicImageSizeMatchModel(image_height, image_width, batch_info)) {
  302. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  303. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  304. }
  305. ret = GraphExecutor::SetDynamicSize(model_id, batch_num, static_cast<int32_t>(DYNAMIC_IMAGE));
  306. if (ret != SUCCESS) {
  307. GELOGE(ret, "Set dynamic size failed");
  308. return ret;
  309. }
  310. // Memcpy dynamic resolution height from host to device
  311. rtError_t rt_ret =
  312. rtMemcpy(dynamic_input_addr, size, &image_height, size, RT_MEMCPY_HOST_TO_DEVICE);
  313. if (rt_ret != RT_ERROR_NONE) {
  314. GELOGE(rt_ret, "memcpy dynamic resolution input data failed! ret: 0x%X", rt_ret);
  315. return RT_ERROR_TO_GE_STATUS(rt_ret);
  316. }
  317. uint64_t remain_size = length - size;
  318. // Memcpy dynamic resolution width from host to device
  319. rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size),
  320. remain_size, &image_width, size, RT_MEMCPY_HOST_TO_DEVICE);
  321. if (rt_ret != RT_ERROR_NONE) {
  322. GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
  323. return RT_ERROR_TO_GE_STATUS(rt_ret);
  324. }
  325. return SUCCESS;
  326. }
  327. Status GeExecutor::SetDynamicDims(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  328. const vector<uint64_t> &dynamic_dims) {
  329. if (dynamic_input_addr == nullptr) {
  330. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic input addr is nullptr!");
  331. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  332. }
  333. vector<uint64_t> cur_dynamic_dims;
  334. Status ret = GetCurDynamicDims(model_id, dynamic_dims, cur_dynamic_dims);
  335. if (ret != SUCCESS) {
  336. GELOGE(ret, "Set cur gear dynamic dims failed");
  337. return ret;
  338. }
  339. std::vector<std::vector<int64_t>> batch_info;
  340. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  341. ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  342. if (ret != SUCCESS) {
  343. GELOGE(ret, "Get dynamic input info failed.");
  344. return ret;
  345. }
  346. if (!IsDynmaicDimsSizeMatchModel(cur_dynamic_dims, batch_info)) {
  347. GELOGE(ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID, "The current dynamic input does not match the gear of the model.");
  348. return ACL_ERROR_GE_DYNAMIC_BATCH_SIZE_INVALID;
  349. }
  350. ret = GraphExecutor::SetDynamicSize(model_id, cur_dynamic_dims, static_cast<int32_t>(DYNAMIC_DIMS));
  351. if (ret != SUCCESS) {
  352. GELOGE(ret, "Set dynamic size failed");
  353. return ret;
  354. }
  355. size_t dynamic_dim_num = cur_dynamic_dims.size();
  356. uint64_t dynamic_input_size = static_cast<uint64_t>(dynamic_dim_num * sizeof(uint32_t));
  357. if (length < dynamic_input_size) {
  358. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  359. "Dynamic input size [%lu] is less than [%lu]!", length, dynamic_input_size);
  360. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  361. }
  362. uint64_t size = sizeof(uint32_t);
  363. if (length >= dynamic_dim_num * sizeof(uint64_t)) {
  364. size = sizeof(uint64_t);
  365. }
  366. rtError_t rt_ret;
  367. for (uint32_t i = 0; i < dynamic_dim_num; ++i) {
  368. // Memcpy dynamic dim[i] from host to device
  369. rt_ret = rtMemcpy(reinterpret_cast<void *>(reinterpret_cast<uint8_t *>(dynamic_input_addr) + size * i),
  370. length - size * i, &cur_dynamic_dims[i], size, RT_MEMCPY_HOST_TO_DEVICE);
  371. if (rt_ret != RT_ERROR_NONE) {
  372. GELOGE(rt_ret, "memcpy dynamic resolution input data failed!");
  373. return RT_ERROR_TO_GE_STATUS(rt_ret);
  374. }
  375. }
  376. return SUCCESS;
  377. }
  378. Status GeExecutor::GetCurDynamicDims(uint32_t model_id, const vector<uint64_t> &dynamic_dims,
  379. vector<uint64_t> &cur_dynamic_dims) {
  380. cur_dynamic_dims.clear();
  381. vector<ge::TensorDesc> input_desc;
  382. vector<ge::TensorDesc> output_desc;
  383. auto ret = GetModelDescInfo(model_id, input_desc, output_desc);
  384. if (ret != ge::SUCCESS) {
  385. GELOGE(ret, "GetModelDescInfo failed.");
  386. return ret;
  387. }
  388. vector<string> user_designate_shape_order;
  389. vector<int64_t> all_data_dims;
  390. ret = GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  391. if (ret != ge::SUCCESS) {
  392. GELOGE(ret, "GetUserDesignateShapeOrder failed.");
  393. return ret;
  394. }
  395. for (auto &data_name : user_designate_shape_order) {
  396. for (auto &desc : input_desc) {
  397. if (desc.GetName() == data_name) {
  398. for (auto dim : desc.GetShape().GetDims()) {
  399. all_data_dims.push_back(dim);
  400. }
  401. break;
  402. }
  403. }
  404. }
  405. if (dynamic_dims.size() != all_data_dims.size()){
  406. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  407. "Dynamic input size [%lu] is not equal with all data dims size [%lu]!",
  408. dynamic_dims.size(), all_data_dims.size());
  409. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  410. }
  411. for (std::size_t i = 0; i < all_data_dims.size(); ++i) {
  412. if (all_data_dims[i] < 0) {
  413. cur_dynamic_dims.push_back(dynamic_dims[i]);
  414. } else if (static_cast<uint64_t>(all_data_dims[i]) != dynamic_dims[i]) {
  415. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID, "Static dims should be same, index: %zu value: %lu should be %ld",
  416. i, dynamic_dims[i], all_data_dims[i]);
  417. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  418. }
  419. }
  420. return SUCCESS;
  421. }
  422. Status GeExecutor::GetCurShape(const uint32_t model_id, std::vector<int64_t> &batch_info, int32_t &dynamic_type) {
  423. GELOGI("Begin to get current shape");
  424. if (!isInit_) {
  425. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  426. return ACL_ERROR_GE_EXEC_NOT_INIT;
  427. }
  428. Status ret = GraphExecutor::GetCurShape(model_id, batch_info, dynamic_type);
  429. if (ret != SUCCESS) {
  430. GELOGE(ret, "Get current shape failed");
  431. return ret;
  432. }
  433. return SUCCESS;
  434. }
  435. Status GeExecutor::SetDynamicAippData(uint32_t model_id, void *dynamic_input_addr, uint64_t length,
  436. const std::vector<kAippDynamicBatchPara> &aippBatchPara,
  437. const kAippDynamicPara &aippParms) {
  438. GELOGI("Enter to SetDynamicAippData.");
  439. if (dynamic_input_addr == nullptr) {
  440. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID, "Dynamic aipp input addr is nullptr!");
  441. return ACL_ERROR_GE_DYNAMIC_INPUT_ADDR_INVALID;
  442. }
  443. if (aippBatchPara.empty()) {
  444. GELOGE(ACL_ERROR_GE_AIPP_BATCH_EMPTY, "aippBatchPara is empty.");
  445. return ACL_ERROR_GE_AIPP_BATCH_EMPTY;
  446. }
  447. uint64_t batch_num = aippBatchPara.size();
  448. uint64_t real_aippParms_size = sizeof(kAippDynamicPara) - sizeof(kAippDynamicBatchPara);
  449. uint64_t struct_len = batch_num * sizeof(kAippDynamicBatchPara) + real_aippParms_size;
  450. GELOGI(
  451. "Get acl input dynamic aipp data, model_id is %u, length is %lu,"
  452. "batch num is %lu, struct_len is %lu",
  453. model_id, length, batch_num, struct_len);
  454. if (struct_len > length) {
  455. GELOGE(ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID,
  456. "input dynamic aipp param len [%lu] is larger than aipp_data size [%lu]", struct_len, length);
  457. return ACL_ERROR_GE_DYNAMIC_INPUT_LENGTH_INVALID;
  458. }
  459. // Memcpy real kAippDynamicBatchPara from host to device
  460. rtError_t rt_ret = rtMemcpy(dynamic_input_addr, length, &aippParms, real_aippParms_size, RT_MEMCPY_HOST_TO_DEVICE);
  461. if (rt_ret != RT_ERROR_NONE) {
  462. GELOGE(rt_ret, "memcpy real_aippParms_size failed! ret: 0x%X", rt_ret);
  463. return RT_ERROR_TO_GE_STATUS(rt_ret);
  464. }
  465. uint64_t remain_len = length - real_aippParms_size;
  466. uint8_t *aipp_batch_para_dev = reinterpret_cast<uint8_t *>(dynamic_input_addr) + real_aippParms_size;
  467. for (uint64_t i = 0; i < batch_num; ++i) {
  468. rt_ret = rtMemcpy(reinterpret_cast<void *>(aipp_batch_para_dev + i * sizeof(kAippDynamicBatchPara)),
  469. (remain_len - i * sizeof(kAippDynamicBatchPara)), &(aippBatchPara[i]),
  470. sizeof(kAippDynamicBatchPara), RT_MEMCPY_HOST_TO_DEVICE);
  471. if (rt_ret != RT_ERROR_NONE) {
  472. GELOGE(rt_ret, "memcpy kAippDynamicBatchPara input data failed! ret: 0x%X", rt_ret);
  473. return RT_ERROR_TO_GE_STATUS(rt_ret);
  474. }
  475. }
  476. return SUCCESS;
  477. }
  478. Status GeExecutor::UnloadModel(uint32_t model_id) {
  479. GELOGD("unload model %u begin.", model_id);
  480. if (!isInit_) {
  481. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  482. return ACL_ERROR_GE_EXEC_NOT_INIT;
  483. }
  484. Status ret = GraphLoader::DestroyAicpuSessionForInfer(model_id);
  485. if (ret != SUCCESS) {
  486. GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
  487. return ret;
  488. }
  489. std::shared_ptr<hybrid::HybridDavinciModel> hybrid_davinci_model =
  490. ModelManager::GetInstance()->GetHybridModel(model_id);
  491. if (hybrid_davinci_model != nullptr) {
  492. uint64_t session_id = hybrid_davinci_model->GetSessionId();
  493. VarManagerPool::Instance().RemoveVarManager(session_id);
  494. } else {
  495. std::shared_ptr<DavinciModel> davinci_model = ModelManager::GetInstance()->GetModel(model_id);
  496. if (davinci_model != nullptr) {
  497. uint64_t session_id = davinci_model->GetSessionId();
  498. VarManagerPool::Instance().RemoveVarManager(session_id);
  499. }
  500. }
  501. ret = GraphLoader::UnloadModel(model_id);
  502. if (ret != SUCCESS) {
  503. GELOGE(ret, "[GraphLoader] DestroyAicpuSessionForInfer failed. model id: %u", model_id);
  504. return ret;
  505. }
  506. return SUCCESS;
  507. }
  508. // Get input and output descriptor
  509. Status GeExecutor::GetModelDescInfo(uint32_t model_id, std::vector<ge::TensorDesc> &input_desc,
  510. std::vector<ge::TensorDesc> &output_desc, bool new_model_desc) {
  511. if (!isInit_) {
  512. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  513. return ACL_ERROR_GE_EXEC_NOT_INIT;
  514. }
  515. std::vector<InputOutputDescInfo> input_desc_infos;
  516. std::vector<InputOutputDescInfo> output_desc_infos;
  517. std::vector<uint32_t> input_formats;
  518. std::vector<uint32_t> output_formats;
  519. Status ret = GraphExecutor::GetInputOutputDescInfo(model_id, input_desc_infos, output_desc_infos, input_formats,
  520. output_formats, new_model_desc);
  521. if (ret != domi::SUCCESS) {
  522. GELOGE(ret, "GetInputOutputDescInfo failed. ret = %u", ret);
  523. return ACL_ERROR_GE_GET_TENSOR_INFO;
  524. }
  525. if (input_formats.size() != input_desc_infos.size()) {
  526. GELOGE(ACL_ERROR_GE_PARAM_INVALID,
  527. "input_formats size %zu is not equal to input_desc_infos size %zu.",
  528. input_formats.size(), input_desc_infos.size());
  529. return ACL_ERROR_GE_PARAM_INVALID;
  530. }
  531. if (output_formats.size() != output_desc_infos.size()) {
  532. GELOGE(ACL_ERROR_GE_PARAM_INVALID, "output_formats size %zu is not equal to output_desc_infos size %zu.",
  533. output_formats.size(), output_desc_infos.size());
  534. return ACL_ERROR_GE_PARAM_INVALID;
  535. }
  536. // Transfer data to TensorDesc
  537. GetGeTensorDescFromDomiInfo(input_desc, input_desc_infos, input_formats);
  538. GetGeTensorDescFromDomiInfo(output_desc, output_desc_infos, output_formats);
  539. return ge::SUCCESS;
  540. }
  541. ///
  542. /// @ingroup ge
  543. /// @brief Get dynamic batch_info
  544. /// @param [in] model_id
  545. /// @param [out] batch_info
  546. /// @param [out] dynamic_type
  547. /// @return execute result
  548. ///
  549. Status GeExecutor::GetDynamicBatchInfo(uint32_t model_id, std::vector<std::vector<int64_t>> &batch_info,
  550. int32_t &dynamic_type) {
  551. if (!isInit_) {
  552. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  553. return ACL_ERROR_GE_EXEC_NOT_INIT;
  554. }
  555. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  556. if (ret != SUCCESS) {
  557. GELOGE(ret, "GetDynamicBatchInfo failed.");
  558. return ret;
  559. }
  560. return SUCCESS;
  561. }
  562. ///
  563. /// @ingroup ge
  564. /// @brief Get combined dynamic dims info
  565. /// @param [in] model_id
  566. /// @param [out] batch_info
  567. /// @return execute result
  568. ///
  569. Status GeExecutor::GetCombinedDynamicDims(uint32_t model_id, vector<vector<int64_t>> &batch_info) {
  570. GELOGI("Begin to get combined dynamic dims info.");
  571. if (!isInit_) {
  572. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  573. return ACL_ERROR_GE_EXEC_NOT_INIT;
  574. }
  575. Status ret = GraphExecutor::GetCombinedDynamicDims(model_id, batch_info);
  576. if (ret != SUCCESS) {
  577. GELOGE(ret, "GetCombinedDynamicDims failed.");
  578. return ret;
  579. }
  580. GELOGI("Get combined dynamic dims succ.");
  581. return SUCCESS;
  582. }
  583. ///
  584. /// @ingroup ge
  585. /// @brief Get user designeate shape order
  586. /// @param [in] model_id
  587. /// @param [out] user_designate_shape_order
  588. /// @return execute result
  589. ///
  590. Status GeExecutor::GetUserDesignateShapeOrder(uint32_t model_id, vector<string> &user_designate_shape_order) {
  591. if (!isInit_) {
  592. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  593. return ACL_ERROR_GE_EXEC_NOT_INIT;
  594. }
  595. Status ret = GraphExecutor::GetUserDesignateShapeOrder(model_id, user_designate_shape_order);
  596. if (ret != SUCCESS) {
  597. GELOGE(ret, "GetUserDesignateShapeOrder failed.");
  598. return ret;
  599. }
  600. return SUCCESS;
  601. }
  602. ///
  603. /// @ingroup ge
  604. /// @brief Get AIPP input format
  605. /// @param [in] model_id
  606. /// @param [in] index
  607. /// @param [out] input_format
  608. /// @return execute result
  609. ///
  610. Status GeExecutor::GetAIPPInfo(uint32_t model_id, uint32_t index, AippConfigInfo &aipp_info) {
  611. GELOGI("Begin to GetAIPPInfo.");
  612. if (!isInit_) {
  613. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  614. return ACL_ERROR_GE_EXEC_NOT_INIT;
  615. }
  616. Status ret = GraphExecutor::GetAippInfo(model_id, index, aipp_info);
  617. if (ret != SUCCESS) {
  618. GELOGW("GetAIPPInfo is not success.");
  619. return ret;
  620. }
  621. GELOGI("GetAIPPInfo succ.");
  622. return SUCCESS;
  623. }
  624. Status GeExecutor::GetAippType(uint32_t model_id, uint32_t index, InputAippType &type, size_t &aipp_index) {
  625. GELOGI("Begin to get aipp type.");
  626. if (!isInit_) {
  627. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  628. return ACL_ERROR_GE_EXEC_NOT_INIT;
  629. }
  630. Status ret = GraphExecutor::GetAippType(model_id, index, type, aipp_index);
  631. if (ret != SUCCESS) {
  632. GELOGW("Get aipp type is not success.");
  633. return ret;
  634. }
  635. GELOGI("Get aipp type success.");
  636. return SUCCESS;
  637. }
  638. Status GeExecutor::GetModelAttr(uint32_t model_id, std::vector<std::string> &dynamic_output_shape_info) {
  639. if (!isInit_) {
  640. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  641. return ACL_ERROR_GE_EXEC_NOT_INIT;
  642. }
  643. Status ret = GraphExecutor::GetModelAttr(model_id, dynamic_output_shape_info);
  644. if (ret != SUCCESS) {
  645. GELOGE(ret, "Get dynamic batch output shape info failed.");
  646. return ret;
  647. }
  648. return SUCCESS;
  649. }
  650. Status GeExecutor::CommandHandle(const Command &command) {
  651. Status ret = GraphLoader::CommandHandle(command);
  652. if (ret != SUCCESS) {
  653. GELOGE(ACL_ERROR_GE_COMMAND_HANDLE, "CommandHandle: Command Handle failed.");
  654. return ACL_ERROR_GE_COMMAND_HANDLE;
  655. }
  656. return SUCCESS;
  657. }
  658. Status GeExecutor::GetMaxUsedMemory(uint32_t model_id, uint32_t &max_size) {
  659. GELOGI("Get max used memory begin.");
  660. if (!isInit_) {
  661. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  662. return ACL_ERROR_GE_EXEC_NOT_INIT;
  663. }
  664. uint64_t max_mem_size = 0;
  665. Status ret = GraphLoader::GetMaxUsedMemory(model_id, max_mem_size);
  666. max_size = static_cast<uint32_t>(max_mem_size);
  667. return ret;
  668. }
  669. /**
  670. * @ingroup ge
  671. * @brief Load data from model file to memory
  672. * @param [in] const std::string &path: Offline model file path
  673. * @param [out] domi::ModelData &model_data: Offline model memory data
  674. * @return SUCCESS handle successfully / others handle failed
  675. */
  676. Status GeExecutor::LoadDataFromFile(const std::string &path, ModelData &model_data) {
  677. GELOGI("Load data from file begin.");
  678. if (!isInit_) {
  679. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  680. return ACL_ERROR_GE_EXEC_NOT_INIT;
  681. }
  682. string filePath = RealPath(path.c_str());
  683. if (filePath.empty()) {
  684. GELOGE(ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID,
  685. "File path is invalid. please check your text file '%s'.", path.c_str());
  686. return ACL_ERROR_GE_EXEC_MODEL_PATH_INVALID;
  687. }
  688. GELOGI("load modelData from file: %s.", path.c_str());
  689. std::string key_path;
  690. int32_t priority = 0;
  691. Status ret = GraphLoader::LoadDataFromFile(path, key_path, priority, model_data);
  692. if (ret != SUCCESS) {
  693. if (model_data.model_data != nullptr) {
  694. delete[] static_cast<char *>(model_data.model_data);
  695. model_data.model_data = nullptr;
  696. }
  697. }
  698. return ret;
  699. }
  700. /**
  701. * @ingroup ge
  702. * @brief Load model from offline model memory data
  703. * @param [in] domi::ModelData &model_data: Offline model data
  704. void *dev_ptr: Input/Output memory start address
  705. size_t memsize: Input/Output memory length
  706. void *weight_ptr: Weight memory start address
  707. size_t weightsize: Weight memory length
  708. * @param [out] uint32_t &model_id: identification after model loading
  709. * @return SUCCESS handle successfully / others handle failed
  710. */
  711. Status GeExecutor::LoadModelFromData(uint32_t &model_id, const ModelData &model_data, void *dev_ptr, size_t mem_size,
  712. void *weight_ptr, size_t weight_size) {
  713. if (!isInit_) {
  714. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  715. return ACL_ERROR_GE_EXEC_NOT_INIT;
  716. }
  717. return GraphLoader::LoadModelFromData(model_id, model_data, dev_ptr, mem_size, weight_ptr, weight_size);
  718. }
  719. /**
  720. * @ingroup ge
  721. * @brief Load task list from ModelData with queue.
  722. * @param [out] model_id: model id allocate from manager.
  723. * @param [in] ge_model_data: Model data load from offline model.
  724. * @param [in] input_queue_ids: input queue ids create from user.
  725. * @param [in] output_queue_ids: input queue ids create from user.
  726. * @return: 0 for success / others for fail
  727. */
  728. Status GeExecutor::LoadModelWithQ(uint32_t &model_id, const ModelData &model_data,
  729. const std::vector<uint32_t> &input_queue_ids,
  730. const std::vector<uint32_t> &output_queue_ids) {
  731. GELOGI("Load model with queue begin.");
  732. if (!isInit_) {
  733. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  734. return ACL_ERROR_GE_EXEC_NOT_INIT;
  735. }
  736. return GraphLoader::LoadModelWithQ(model_id, model_data, input_queue_ids, output_queue_ids);
  737. }
  738. /**
  739. * @ingroup ge
  740. * @brief Synchronous execution of offline model(Do not create thread)
  741. * @param [in] uint32_t model_id: Model ID to execute
  742. void* stream: stream to execute
  743. const domi::InputData *input_data: Model input data
  744. bool async_mode: is asynchronize mode.
  745. * @param [out] domi::OutputData *output_data: Model output data
  746. * @return SUCCESS handle successfully / others handle failed
  747. */
  748. Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
  749. ge::RunModelData &run_output_data, bool async_mode) {
  750. std::vector<GeTensorDesc> input_desc = {};
  751. std::vector<GeTensorDesc> output_desc = {};
  752. return ExecModel(model_id, stream, run_input_data, input_desc, run_output_data, output_desc, async_mode);
  753. }
  754. /**
  755. * @ingroup ge
  756. * @brief Synchronous execution of offline model(Do not create thread)
  757. * @param [in] uint32_t model_id: Model ID to execute
  758. void* stream: stream to execute
  759. const domi::InputData *input_data: Model input data
  760. const std::vector<GeTensorDesc> &input_desc: Description of model input data
  761. bool async_mode: is asynchronize mode
  762. * @param [out] domi::OutputData *output_data: Model output data
  763. * @param [out] std::vector<GeTensorDesc> &output_desc: Description of model output data
  764. * @return SUCCESS handle successfully / others handle failed
  765. */
  766. Status GeExecutor::ExecModel(uint32_t model_id, void *stream, const ge::RunModelData &run_input_data,
  767. const std::vector<GeTensorDesc> &input_desc, ge::RunModelData &run_output_data,
  768. std::vector<GeTensorDesc> &output_desc, bool async_mode) {
  769. if (!isInit_) {
  770. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  771. return ACL_ERROR_GE_EXEC_NOT_INIT;
  772. }
  773. InputData input_data;
  774. OutputData output_data;
  775. GetDomiInputData(run_input_data, input_data);
  776. GetDomiOutputData(run_output_data, output_data);
  777. if ((run_input_data.dynamic_batch_size != 0) || (run_input_data.dynamic_image_width != 0) ||
  778. (run_input_data.dynamic_image_height != 0) || (run_input_data.dynamic_dims.size() != 0)) {
  779. std::vector<std::vector<int64_t>> batch_info;
  780. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  781. Status ret = GraphExecutor::GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  782. if (ret != SUCCESS) {
  783. GELOGE(ret, "Get dynamic input info failed.");
  784. return ret;
  785. }
  786. if (!batch_info.empty()) {
  787. SetDynamicInputDataFlag(run_input_data, batch_info, input_data);
  788. }
  789. }
  790. return GraphLoader::ExecuteModel(model_id, stream, async_mode, input_data, input_desc, output_data, output_desc);
  791. }
  792. /**
  793. * @ingroup ge
  794. * @brief Get weight memory size from model file
  795. * @param [in] const std::string &path: Offline model file path
  796. * @param [out] size_t &mem_size Execution memory size
  797. size_t &weight_size Weight memory space size
  798. * @return SUCCESS handle successfully / others handle failed
  799. */
  800. Status GeExecutor::GetMemAndWeightSize(const std::string &path, size_t &mem_size, size_t &weight_size) {
  801. GELOGI("Get memory and weight size from file begin.");
  802. if (!isInit_) {
  803. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  804. return ACL_ERROR_GE_EXEC_NOT_INIT;
  805. }
  806. ModelData model;
  807. std::string key;
  808. Status ret = ge::GraphLoader::LoadDataFromFile(path, key, 0, model);
  809. if ((ret != SUCCESS) || (model.model_data == nullptr)) {
  810. GELOGE(ret, "Load data from file failed. ret = %d", ret);
  811. return ret;
  812. }
  813. ret = ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  814. delete[] static_cast<char *>(model.model_data);
  815. model.model_data = nullptr;
  816. return ret;
  817. }
  818. /**
  819. * @ingroup ge
  820. * @brief Get weight memory size from model file
  821. * @param [in] const void *model_data Offline model buffer
  822. size_t model_size Offline model buffer length
  823. * @param [out] size_t &mem_size Execution memory size
  824. size_t &weight_size Weight memory space size
  825. * @return SUCCESS handle successfully / others handle failed
  826. */
  827. Status GeExecutor::GetMemAndWeightSize(const void *model_data, size_t model_size, size_t &mem_size,
  828. size_t &weight_size) {
  829. GELOGI("Get memory and weight size from data begin.");
  830. if (!isInit_) {
  831. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "GeExecutor has not been initialized!");
  832. return ACL_ERROR_GE_EXEC_NOT_INIT;
  833. }
  834. if (model_data == nullptr) {
  835. GELOGE(ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID, "invalid model data!");
  836. return ACL_ERROR_GE_EXEC_MODEL_ADDR_INVALID;
  837. }
  838. ModelData model;
  839. model.model_data = const_cast<void *>(model_data);
  840. model.model_len = static_cast<uint32_t>(model_size);
  841. return ge::ModelManager::GetModelMemAndWeightSize(model, mem_size, weight_size);
  842. }
  843. Status GeExecutor::LoadSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  844. SingleOp **single_op) {
  845. return SingleOpManager::GetInstance().GetOpFromModel(model_name, modelData, stream, single_op);
  846. }
  847. Status GeExecutor::LoadDynamicSingleOp(const std::string &model_name, const ge::ModelData &modelData, void *stream,
  848. DynamicSingleOp **single_op) {
  849. return SingleOpManager::GetInstance().GetDynamicOpFromModel(model_name, modelData, stream, single_op);
  850. }
  851. Status GeExecutor::ExecuteAsync(SingleOp *executor, const std::vector<DataBuffer> &inputs,
  852. std::vector<DataBuffer> &outputs) {
  853. if (executor == nullptr) {
  854. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "param is NULL");
  855. return ACL_ERROR_GE_EXEC_NOT_INIT;
  856. }
  857. return executor->ExecuteAsync(inputs, outputs);
  858. }
  859. ge::Status GeExecutor::ExecuteAsync(DynamicSingleOp *executor, const vector<GeTensorDesc> &input_desc,
  860. const vector<DataBuffer> &inputs, vector<GeTensorDesc> &output_desc,
  861. vector<DataBuffer> &outputs) {
  862. GE_CHECK_NOTNULL(executor);
  863. return executor->ExecuteAsync(input_desc, inputs, output_desc, outputs);
  864. }
  865. Status GeExecutor::ReleaseSingleOpResource(void *stream) {
  866. ModelManager::GetInstance()->ClearAicpuSo();
  867. return SingleOpManager::GetInstance().ReleaseResource(stream);
  868. }
  869. Status GeExecutor::GetDeviceIdByModelId(uint32_t model_id, uint32_t &device_id) {
  870. auto model_manager = ModelManager::GetInstance();
  871. GE_CHECK_NOTNULL(model_manager);
  872. auto davinci_model = model_manager->GetModel(model_id);
  873. if (davinci_model == nullptr) {
  874. GELOGE(ACL_ERROR_GE_EXEC_MODEL_ID_INVALID, "Model id: %d is invaild or model is not loaded.", model_id);
  875. return ACL_ERROR_GE_EXEC_MODEL_ID_INVALID;
  876. }
  877. device_id = davinci_model->GetDeviceId();
  878. return SUCCESS;
  879. }
  880. Status GeExecutor::GetBatchInfoSize(uint32_t model_id, size_t &shape_count) {
  881. std::vector<std::vector<int64_t>> batch_info;
  882. int32_t dynamic_type = static_cast<int32_t>(FIXED);
  883. Status ret = GetDynamicBatchInfo(model_id, batch_info, dynamic_type);
  884. if (ret != SUCCESS) {
  885. GELOGE(ret, "Calc batch info size failed. ret = %d", ret);
  886. return ret;
  887. }
  888. if (batch_info.empty()) {
  889. shape_count = kStaticBatchInfoSize;
  890. } else {
  891. shape_count = batch_info.size();
  892. }
  893. return SUCCESS;
  894. }
  895. Status GeExecutor::GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info) {
  896. GELOGI("Begin to GetOrigInputInfo.");
  897. if (!isInit_) {
  898. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  899. return ACL_ERROR_GE_EXEC_NOT_INIT;
  900. }
  901. Status ret = GraphExecutor::GetOrigInputInfo(model_id, index, orig_input_info);
  902. if (ret != SUCCESS) {
  903. GELOGE(ret, "GetOrigInputInfo failed.");
  904. return ret;
  905. }
  906. GELOGI("GetOrigInputInfo succ.");
  907. return SUCCESS;
  908. }
  909. Status GeExecutor::GetAllAippInputOutputDims(uint32_t model_id, uint32_t index,
  910. std::vector<InputOutputDims> &input_dims,
  911. std::vector<InputOutputDims> &output_dims) {
  912. GELOGI("Begin to GetAllAippInputOutputDims.");
  913. if (!isInit_) {
  914. GELOGE(ACL_ERROR_GE_EXEC_NOT_INIT, "not inited yet!");
  915. return ACL_ERROR_GE_EXEC_NOT_INIT;
  916. }
  917. Status ret = GraphExecutor::GetAllAippInputOutputDims(model_id, index, input_dims, output_dims);
  918. if (ret != SUCCESS) {
  919. GELOGE(ret, "GetAllAippInputOutputDims failed.");
  920. return ret;
  921. }
  922. GELOGI("GetAllAippInputOutputDims succ.");
  923. return SUCCESS;
  924. }
  925. Status GeExecutor::GetOpDescInfo(uint32_t device_id, uint32_t stream_id, uint32_t task_id, OpDescInfo &op_desc_info) {
  926. GELOGI("Begin to GetOpDescInfo.");
  927. Status ret = GraphExecutor::GetOpDescInfo(device_id, stream_id, task_id, op_desc_info);
  928. if (ret != SUCCESS) {
  929. GELOGE(ret, "GetOpDescInfo failed.");
  930. return ret;
  931. }
  932. GELOGI("GetOpDescInfo succ.");
  933. return SUCCESS;
  934. }
  935. Status GeExecutor::SetDump(const DumpConfig &dump_config) {
  936. GELOGI("Start to set dump config");
  937. auto ret = DumpManager::GetInstance().SetDumpConf(dump_config);
  938. if (ret != SUCCESS) {
  939. GELOGE(ret, "Set dump conf failed");
  940. return ret;
  941. }
  942. GELOGI("Set dump config successfully");
  943. return SUCCESS;
  944. }
  945. } // namespace ge

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示