You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

imperative_rt.cpp 4.4 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107
  1. /**
  2. * \file imperative/python/src/imperative_rt.cpp
  3. * MegEngine is Licensed under the Apache License, Version 2.0 (the "License")
  4. *
  5. * Copyright (c) 2014-2020 Megvii Inc. All rights reserved.
  6. *
  7. * Unless required by applicable law or agreed to in writing,
  8. * software distributed under the License is distributed on an
  9. * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  10. */
  11. #include "./imperative_rt.h"
  12. #include <future>
  13. #include <variant>
  14. #include <unordered_map>
  15. #include <pybind11/numpy.h>
  16. #include <pybind11/operators.h>
  17. #include "megbrain/imperative.h"
  18. #include "megbrain/imperative/interpreter.h"
  19. #include "megbrain/imperative/ops/opr_attr.h"
  20. #include "./helper.h"
  21. namespace py = pybind11;
  22. using namespace mgb;
  23. using namespace imperative;
  24. using namespace interpreter;
  25. namespace {
  26. std::optional<std::tuple<std::shared_ptr<OpDef>, std::vector<bool>, std::vector<bool>>>
  27. make_backward_graph(
  28. const OpDef& opdef, std::vector<LogicalTensorDesc> inputs,
  29. std::vector<bool> input_requires_grad,
  30. std::vector<bool> output_has_grad) {
  31. auto res = OpDef::make_backward_graph(opdef,
  32. SmallVector<LogicalTensorDesc>(inputs.begin(), inputs.end()),
  33. SmallVector<bool>(input_requires_grad.begin(), input_requires_grad.end()),
  34. SmallVector<bool>(output_has_grad.begin(), output_has_grad.end()));
  35. if (res.backward) {
  36. return std::optional<std::tuple<std::shared_ptr<OpDef>, std::vector<bool>, std::vector<bool>>>{
  37. std::in_place, res.backward, res.save_for_backward, res.input_has_grad};
  38. } else {
  39. return {};
  40. }
  41. }
  42. } // namespace
  43. void init_imperative_rt(py::module m) {
  44. py::class_<Interpreter::Channel>(m, "Interpreter")
  45. .def("put", [](Interpreter::Channel& self, py::array data, DType dtype, CompNode cn) {
  46. if (!cn.valid()) {
  47. cn = CompNode::load("xpux");
  48. }
  49. constexpr int size_threshhold = TensorShape::MAX_NDIM;
  50. if (data.size() > size_threshhold) {
  51. return self.put(npy::np2tensor(data.ptr(), npy::Meth::borrow(cn), dtype));
  52. } else {
  53. HostTensorND ret(cn);
  54. return self.put(npy::np2tensor(data.ptr(), npy::Meth::copy_into(&ret), dtype));
  55. }
  56. }, py::arg(), py::arg("dtype") = py::none(), py::arg("device") = py::none())
  57. .def("put", py::overload_cast<const DeviceTensorND&>(&Interpreter::Channel::put))
  58. .def("delete", [](Interpreter::Channel& self, Interpreter::Handle handle) {
  59. return self.del(handle);
  60. })
  61. .def("get_value", [](Interpreter::Channel& self, Interpreter::Handle handle) {
  62. PyObject* optr = npy::ndarray_from_tensor(self.get_value(handle), npy::ShareType::TRY_SHARE);
  63. return py::reinterpret_steal<py::object>(optr);
  64. })
  65. .def("get_dtype", &Interpreter::Channel::get_dtype)
  66. .def("get_device", &Interpreter::Channel::get_device)
  67. .def("get_shape", &Interpreter::Channel::get_shape)
  68. .def("_get_dev_tensor", &Interpreter::Channel::get_dev_tensor)
  69. .def("apply_op", &Interpreter::Channel::apply_op)
  70. .def("sync", &Interpreter::Channel::sync, py::call_guard<py::gil_scoped_release>());
  71. std::unique_ptr<Interpreter::Channel> ch = Interpreter::inst().create_channel();
  72. m.attr("interpreter") = py::detail::make_caster<decltype(ch)>::cast(
  73. std::move(ch), py::return_value_policy::move, {});
  74. for (auto name : {"put", "delete", "get_value", "get_dtype", "get_device", "get_shape", "_get_dev_tensor", "apply_op"}) {
  75. m.attr(name) = m.attr("interpreter").attr(name);
  76. }
  77. m.def("sync", [m]() {
  78. m.attr("interpreter").attr("sync")();
  79. py::gil_scoped_release _;
  80. py_task_q.wait_all_task_finish();
  81. });
  82. m.def("make_backward_graph", &make_backward_graph);
  83. py::class_<OpDef, std::shared_ptr<OpDef>>(m, "OpDef")
  84. .def("ctype", [](const OpDef& opdef) {
  85. if (auto attr = opdef.try_cast_final<OprAttr>()) {
  86. return attr->type.c_str();
  87. }
  88. return opdef.dyn_typeinfo()->name;
  89. })
  90. .def("__eq__", [](const OpDef& lhs, const OpDef& rhs) {
  91. return lhs.is_same(rhs);
  92. })
  93. .def("__hash__", &OpDef::hash);
  94. }

MegEngine 安装包中集成了使用 GPU 运行代码所需的 CUDA 环境,不用区分 CPU 和 GPU 版。 如果想要运行 GPU 程序,请确保机器本身配有 GPU 硬件设备并安装好驱动。 如果你想体验在云端 GPU 算力平台进行深度学习开发的感觉,欢迎访问 MegStudio 平台