|
|
@@ -65,23 +65,6 @@ class AicpuNodeTaskBase : public NodeTask { |
|
|
|
|
|
|
|
static Status AllocTensorBuffer(size_t size, std::unique_ptr<TensorBuffer> &tensor_buffer); |
|
|
|
|
|
|
|
virtual Status CopyDataToHbm(TaskContext &context, |
|
|
|
const std::vector<std::unique_ptr<TensorBuffer>> &out_shape_hbm) = 0; |
|
|
|
|
|
|
|
/// |
|
|
|
/// read result summary and prepare copy task memory. |
|
|
|
/// @param context task context |
|
|
|
/// @param out_shape_hbm if scalar, TensorBuffer->data is null, size=0 |
|
|
|
/// @return SUCCESS:success other:failed |
|
|
|
/// |
|
|
|
Status ReadResultSummaryAndPrepareMemory(TaskContext &context, |
|
|
|
std::vector<std::unique_ptr<TensorBuffer>> &out_shape_hbm); |
|
|
|
|
|
|
|
Status UpdateShapeByHbmBuffer(TaskContext &context, |
|
|
|
const std::vector<std::unique_ptr<TensorBuffer>> &out_shape_hbm); |
|
|
|
|
|
|
|
Status PrepareCopyInputs(const TaskContext &context, |
|
|
|
const std::vector<std::unique_ptr<TensorBuffer>> &out_shape_hbm); |
|
|
|
Status DistributeWaitTaskForAicpuBlockingOp(rtStream_t stream); |
|
|
|
Status CheckDeviceSupportBlockingAicpuOpProcess(bool &is_support); |
|
|
|
Status UpdateEventIdForBlockingAicpuOp(); |
|
|
@@ -132,14 +115,6 @@ class AicpuNodeTaskBase : public NodeTask { |
|
|
|
// for blocking aicpu op |
|
|
|
bool is_blocking_aicpu_op_ = false; |
|
|
|
rtEvent_t rt_event_ = nullptr; |
|
|
|
|
|
|
|
std::vector<std::unique_ptr<TensorBuffer>> output_summary_; |
|
|
|
std::vector<aicpu::FWKAdapter::ResultSummary> output_summary_host_; |
|
|
|
|
|
|
|
std::unique_ptr<TensorBuffer> copy_input_release_flag_dev_; |
|
|
|
std::unique_ptr<TensorBuffer> copy_input_data_size_dev_; |
|
|
|
std::unique_ptr<TensorBuffer> copy_input_src_dev_; |
|
|
|
std::unique_ptr<TensorBuffer> copy_input_dst_dev_; |
|
|
|
}; |
|
|
|
|
|
|
|
class AicpuTfNodeTask : public AicpuNodeTaskBase { |
|
|
|