You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

group_conv3d.cpp 6.3 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159
  1. /**
  2. * \file dnn/test/cuda/group_conv3d.cpp
  3. * MegEngine is Licensed under the Apache License, Version 2.0 (the "License")
  4. *
  5. * Copyright (c) 2014-2020 Megvii Inc. All rights reserved.
  6. *
  7. * Unless required by applicable law or agreed to in writing,
  8. * software distributed under the License is distributed on an
  9. * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  10. */
  11. #include "megdnn/oprs/nn.h"
  12. #include "test/common/benchmarker.h"
  13. #include "test/common/checker.h"
  14. #include "test/common/convolution3d.h"
  15. #include "test/cuda/fixture.h"
  16. #include "src/cuda/utils.h"
  17. namespace megdnn {
  18. namespace test {
  19. TEST_F(CUDA, GROUP_CONVOLUTION3D_FORWARD) {
  20. bool is_int_available = cuda::is_compute_capability_required(6, 1);
  21. static_cast<void>(is_int_available);
  22. auto run = [&](size_t N, size_t IC, size_t ID, size_t IH, size_t IW,
  23. size_t FD, size_t FH, size_t FW, size_t OC, size_t PD,
  24. size_t PH, size_t PW, size_t SD, size_t SH, size_t SW,
  25. size_t DD, size_t DH, size_t DW, size_t group) {
  26. {
  27. // float case
  28. Checker<Convolution3D> checker(handle_cuda());
  29. Convolution3D::Param param;
  30. param.sparse = Convolution3D::Param::Sparse::GROUP;
  31. param.pad_d = PD;
  32. param.pad_h = PH;
  33. param.pad_w = PW;
  34. param.stride_d = SD;
  35. param.stride_h = SH;
  36. param.stride_w = SW;
  37. param.dilate_d = DD;
  38. param.dilate_h = DH;
  39. param.dilate_w = DW;
  40. auto ICpg = IC / group;
  41. auto OCpg = OC / group;
  42. checker.set_param(param).exec(
  43. {{N, IC, ID, IH, IW}, {group, OCpg, ICpg, FD, FH, FW}, {}});
  44. }
  45. };
  46. // normal case
  47. run(2, 64, 7, 7, 7, 1, 1, 1, 32, 0, 0, 0, 1, 1, 1, 1, 1, 1, 2);
  48. run(1, 2, 2, 2, 2, 1, 1, 2, 2, 0, 0, 0, 1, 1, 1, 1, 1, 1, 2);
  49. run(2, 64, 7, 7, 7, 3, 3, 3, 32, 0, 0, 0, 1, 1, 1, 1, 1, 1, 2);
  50. // padded case
  51. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 2, 2, 2, 1, 1, 1, 1, 1, 1, 4);
  52. // strided case
  53. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 0, 0, 0, 2, 2, 2, 1, 1, 1, 8);
  54. // dilated case
  55. #if CUDNN_MAJOR >= 6
  56. run(10, 4, 64, 64, 12, 3, 2, 2, 64, 0, 0, 0, 1, 1, 1, 3, 4, 2, 4);
  57. #else
  58. #endif
  59. }
  60. TEST_F(CUDA, GROUP_CONVOLUTION3D_FORWARD_1x1x1) {
  61. auto run = [&](size_t N, size_t IC, size_t ID, size_t IH, size_t IW,
  62. size_t FD, size_t FH, size_t FW, size_t OC, size_t group) {
  63. Checker<Convolution3D> checker(handle_cuda());
  64. #if CUDNN_MAJOR <= 6
  65. bool require_algo = true;
  66. checker.set_before_exec_callback(
  67. AlgoChecker<Convolution3DForward>{
  68. "group_conv3d:1x1x1", &require_algo});
  69. #endif
  70. Convolution3D::Param param;
  71. param.sparse = Convolution3D::Param::Sparse::GROUP;
  72. auto ICg = IC / group;
  73. auto OCg = OC / group;
  74. checker.set_param(param).exec(
  75. {{N, IC, ID, IH, IW}, {group, OCg, ICg, FD, FH, FW}, {}});
  76. };
  77. size_t ic = 192;
  78. for (size_t g = 2; g <= 4; g += 1) {
  79. for (size_t id = 4; id <= 16; id *= 2) {
  80. size_t iw = id, ih = id;
  81. run(2, ic, id, ih, iw, 1, 1, 1, ic / g, g);
  82. run(2, ic, id + 1, ih + 1, iw + 1, 1, 1, 1, ic / g, g);
  83. }
  84. }
  85. }
  86. TEST_F(CUDA, GROUP_CONVOLUTION3D_BACKWARD_DATA) {
  87. auto run = [&](size_t N, size_t IC, size_t ID, size_t IH, size_t IW,
  88. size_t FD, size_t FH, size_t FW, size_t OC, size_t OD,
  89. size_t OH, size_t OW, size_t PD, size_t PH, size_t PW,
  90. size_t SD, size_t SH, size_t SW, size_t group) {
  91. Checker<Convolution3DBackwardData> checker(handle_cuda());
  92. Convolution3DBackwardData::Param param;
  93. param.sparse = Convolution3D::Param::Sparse::GROUP;
  94. param.pad_d = PD;
  95. param.pad_h = PH;
  96. param.pad_w = PW;
  97. param.stride_d = SD;
  98. param.stride_h = SH;
  99. param.stride_w = SW;
  100. auto ICg = IC / group;
  101. auto OCg = OC / group;
  102. checker.set_param(param).exec({{group, OCg, ICg, FD, FH, FW},
  103. {N, OC, OD, OH, OW},
  104. {N, IC, ID, IH, IW}});
  105. };
  106. // bug case in prev ver
  107. run(1, 2, 1, 1, 1, 1, 1, 1, 2, 1, 1, 3, 0, 0, 1, 1, 1, 1, 2);
  108. run(1, 2, 1, 1, 1, 1, 1, 1, 2, 1, 1, 2, 0, 0, 1, 1, 1, 2, 2);
  109. run(1, 2, 1, 1, 1, 1, 1, 1, 2, 1, 2, 1, 0, 1, 0, 1, 2, 1, 2);
  110. run(1, 2, 1, 1, 1, 1, 1, 1, 2, 2, 1, 1, 1, 0, 0, 2, 1, 1, 2);
  111. // normal case
  112. run(2, 64, 7, 7, 7, 3, 3, 3, 32, 5, 5, 5, 0, 0, 0, 1, 1, 1, 2);
  113. // padded case
  114. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 7, 7, 7, 1, 1, 1, 1, 1, 1, 4);
  115. // strided case
  116. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 3, 3, 3, 0, 0, 0, 2, 2, 2, 8);
  117. // bigger case
  118. run(2, 32, 64, 64, 64, 3, 3, 3, 32, 62, 62, 62, 0, 0, 0, 1, 1, 1, 4);
  119. }
  120. TEST_F(CUDA, GROUP_CONVOLUTION3D_BACKWARD_FILTER) {
  121. auto run = [&](size_t N, size_t IC, size_t ID, size_t IH, size_t IW,
  122. size_t FD, size_t FH, size_t FW, size_t OC, size_t OD,
  123. size_t OH, size_t OW, size_t PD, size_t PH, size_t PW,
  124. size_t SD, size_t SH, size_t SW, size_t group) {
  125. Checker<Convolution3DBackwardFilter> checker(handle_cuda());
  126. Convolution3DBackwardFilter::Param param;
  127. param.sparse = Convolution3D::Param::Sparse::GROUP;
  128. param.pad_d = PD;
  129. param.pad_h = PH;
  130. param.pad_w = PW;
  131. param.stride_d = SD;
  132. param.stride_h = SH;
  133. param.stride_w = SW;
  134. auto ICg = IC / group;
  135. auto OCg = OC / group;
  136. checker.set_param(param).exec({{N, IC, ID, IH, IW},
  137. {N, OC, OD, OH, OW},
  138. {group, OCg, ICg, FD, FH, FW}});
  139. };
  140. // normal case
  141. run(2, 64, 7, 7, 7, 3, 3, 3, 32, 5, 5, 5, 0, 0, 0, 1, 1, 1, 2);
  142. // padded case
  143. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 7, 7, 7, 1, 1, 1, 1, 1, 1, 4);
  144. // strided case
  145. run(2, 32, 7, 7, 7, 3, 3, 3, 64, 3, 3, 3, 0, 0, 0, 2, 2, 2, 8);
  146. }
  147. } // namespace test
  148. } // namespace megdnn
  149. // vim: syntax=cpp.doxygen

MegEngine 安装包中集成了使用 GPU 运行代码所需的 CUDA 环境,不用区分 CPU 和 GPU 版。 如果想要运行 GPU 程序,请确保机器本身配有 GPU 硬件设备并安装好驱动。 如果你想体验在云端 GPU 算力平台进行深度学习开发的感觉,欢迎访问 MegStudio 平台