You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

conv_bias_multi_thread_im2col.cpp 23 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576
  1. /**
  2. * \file dnn/test/arm_common/conv_bias_multi_thread_im2col.cpp
  3. * MegEngine is Licensed under the Apache License, Version 2.0 (the "License")
  4. *
  5. * Copyright (c) 2014-2021 Megvii Inc. All rights reserved.
  6. *
  7. * Unless required by applicable law or agreed to in writing,
  8. * software distributed under the License is distributed on an
  9. * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or
  10. * implied.
  11. */
  12. #include "megdnn/dtype.h"
  13. #include "test/arm_common/fixture.h"
  14. #include "test/common/benchmarker.h"
  15. #include "test/common/conv_bias.h"
  16. #include "test/arm_common/cpuinfo_help.h"
  17. using namespace megdnn;
  18. using namespace test;
  19. using namespace conv_bias;
  20. TEST_F(ARM_COMMON_MULTI_THREADS, CONVBIAS_IM2COL_FP32_STRIDE2) {
  21. #define cb(name) \
  22. check_conv_bias( \
  23. get_conv_bias_args({1, 2, 3, 4, 5, 6, 7}, 2, false, false, false), \
  24. handle(), name);
  25. #if MEGDNN_AARCH64
  26. cb("IM2COLMATMUL:AARCH64_F32K8X12X1")
  27. cb("IM2COLMATMUL:AARCH64_F32K4X16X1")
  28. cb("IM2COLMATMUL:FB_F32_K8X12X1")
  29. #elif MEGDNN_ARMV7
  30. cb("IM2COLMATMUL:ARMV7_F32")
  31. #endif
  32. #undef cb
  33. }
  34. TEST_F(ARM_COMMON_MULTI_THREADS, CONVBIAS_IM2COL_FP32_STRIDE1) {
  35. #define cb(name) \
  36. check_conv_bias( \
  37. get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, false, false), \
  38. handle(), name);
  39. #if MEGDNN_AARCH64
  40. cb("IM2COLMATMUL:AARCH64_F32K8X12X1")
  41. cb("IM2COLMATMUL:AARCH64_F32K4X16X1")
  42. cb("IM2COLMATMUL:FB_F32_K8X12X1")
  43. #elif MEGDNN_ARMV7
  44. cb("IM2COLMATMUL:ARMV7_F32")
  45. cb("IM2COLMATMUL:FB_F32_K8X12X1")
  46. #endif
  47. #undef cb
  48. }
  49. //! CPUINFO ralated test
  50. #if MEGDNN_AARCH64
  51. #if MGB_ENABLE_CPUINFO
  52. TEST_F(ARM_COMMON_MULTI_THREADS, CONVBIAS_IM2COL_FP32_A55) {
  53. CpuInfoTmpReplace cpu_replace_guard(cpuinfo_uarch_cortex_a55);
  54. #define cb(name,stride) \
  55. check_conv_bias( \
  56. get_conv_bias_args({2, 3, 4, 5, 6, 7}, stride, false, false, false), \
  57. handle(), name);
  58. cb("IM2COLMATMUL:AARCH64_F32K8X12X1", 1)
  59. cb("IM2COLMATMUL:AARCH64_F32K8X12X1", 2)
  60. #undef cb
  61. }
  62. #endif
  63. #endif
  64. #if MEGDNN_AARCH64
  65. #if MGB_ENABLE_CPUINFO
  66. TEST_F(ARM_COMMON_MULTI_THREADS, CONVBIAS_IM2COL_FP32_A53) {
  67. CpuInfoTmpReplace cpu_replace_guard(cpuinfo_uarch_cortex_a53);
  68. #define cb(name,stride) \
  69. check_conv_bias( \
  70. get_conv_bias_args({2, 3, 4, 5, 6, 7}, stride, false, false, false), \
  71. handle(), name);
  72. cb("IM2COLMATMUL:AARCH64_F32K8X12X1", 1)
  73. cb("IM2COLMATMUL:AARCH64_F32K8X12X1", 2)
  74. #undef cb
  75. }
  76. #endif
  77. #endif
  78. #if MEGDNN_AARCH64
  79. #if MGB_ENABLE_CPUINFO
  80. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COL_MK4_PACK_F32_A55) {
  81. CpuInfoTmpReplace cpu_replace_guard(cpuinfo_uarch_cortex_a55);
  82. using namespace conv_bias;
  83. std::vector<conv_bias::TestArg> args =
  84. get_nchw44_conv_bias_args({2,3,7},FULL_NLMODE,ONLY_NO_BIASMODE,1);
  85. check_conv_bias(args, handle(), "IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  86. args = get_nchw44_conv_bias_args({2,3,7},FULL_NLMODE,ONLY_NO_BIASMODE,2);
  87. check_conv_bias(args, handle(), "IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  88. }
  89. #endif
  90. #endif
  91. #if MEGDNN_AARCH64
  92. #if MGB_ENABLE_CPUINFO
  93. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COL_MK4_PACK_F32_A53) {
  94. CpuInfoTmpReplace cpu_replace_guard(cpuinfo_uarch_cortex_a53);
  95. using namespace conv_bias;
  96. std::vector<conv_bias::TestArg> args =
  97. get_nchw44_conv_bias_args({2,3,7},FULL_NLMODE,ONLY_NO_BIASMODE,1);
  98. check_conv_bias(args, handle(), "IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  99. args = get_nchw44_conv_bias_args({2,3,7},FULL_NLMODE,ONLY_NO_BIASMODE,2);
  100. check_conv_bias(args, handle(), "IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  101. }
  102. #endif
  103. #endif
  104. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM) {
  105. UniformIntRNG rng{-50, 50};
  106. #define cb(name) \
  107. checker_conv_bias_common( \
  108. get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, false, false, \
  109. true, true), \
  110. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  111. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  112. dtype::QuantizedS8(60.25f), name); \
  113. checker_conv_bias_common( \
  114. get_conv_bias_args({1}, 2, false, false, false, true, true), \
  115. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  116. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  117. dtype::QuantizedS8(60.25f), name);
  118. float epsilon = 0.001;
  119. #if MEGDNN_AARCH64
  120. #if MGB_ENABLE_DOT
  121. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K8X12X4_DOTPROD");
  122. #else
  123. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K8X8X8");
  124. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K4X4X16");
  125. #endif
  126. #elif MEGDNN_ARMV7
  127. epsilon = 1;
  128. cb("IM2COLMATMUL:ARMV7_INT8X8X32_K4X8X8");
  129. #endif
  130. #undef cb
  131. }
  132. #if MGB_ENABLE_DOT
  133. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_MK4_DOT) {
  134. UniformIntRNG rng{-50, 50};
  135. #define cb(name) \
  136. checker_conv_bias_common( \
  137. get_nchw44_conv_bias_args({2, 3, 4, 5, 6, 7}, QUAN_NLMODE, \
  138. BR_AND_NO_BIASMODE, 1, false, false, \
  139. true), \
  140. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  141. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  142. dtype::QuantizedS8(60.25f), name); \
  143. checker_conv_bias_common( \
  144. get_nchw44_conv_bias_args({1}, ONLY_IDENTITY_NLMODE, \
  145. ONLY_BR_BIASMODE, 2, false, false, \
  146. true), \
  147. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  148. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  149. dtype::QuantizedS8(60.25f), name);
  150. float epsilon = 0.001;
  151. #if MEGDNN_AARCH64
  152. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_8X12X4_DOTPROD:96");
  153. #elif MEGDNN_ARMV7
  154. epsilon = 1;
  155. cb("IM2COLMATMUL:AARCH32_INT8_MK4_8X4X4_DOTPROD:96");
  156. #endif
  157. #undef cb
  158. }
  159. TEST_F(ARM_COMMON_MULTI_THREADS,
  160. CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_MK4_DOT_S2_FUSE) {
  161. UniformIntRNG rng{-50, 50};
  162. #define cb(name) \
  163. checker_conv_bias_common( \
  164. get_nchw44_conv_bias_args({3}, QUAN_NLMODE, BR_AND_NO_BIASMODE, 2, \
  165. false, false, true), \
  166. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  167. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  168. dtype::QuantizedS8(60.25f), name);
  169. float epsilon = 0.001;
  170. #if MEGDNN_AARCH64
  171. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_8X12X4_DOTPROD:96");
  172. #elif MEGDNN_ARMV7
  173. epsilon = 1;
  174. cb("IM2COLMATMUL:AARCH32_INT8_MK4_8X4X4_DOTPROD:96");
  175. #endif
  176. #undef cb
  177. }
  178. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_S8x8x32_MK4_DOT) {
  179. UniformIntRNG rng{-50, 50};
  180. #define cb(name) \
  181. checker_conv_bias_common( \
  182. get_nchw44_conv_bias_args( \
  183. {2, 3, 4, 5, 6, 7}, ONLY_IDENTITY_NLMODE, \
  184. BR_AND_BIAS_BIASMODE, 1, false, false, true), \
  185. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  186. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), {}, name); \
  187. checker_conv_bias_common( \
  188. get_nchw44_conv_bias_args({1}, ONLY_IDENTITY_NLMODE, \
  189. BR_AND_BIAS_BIASMODE, 2, false, false, \
  190. true), \
  191. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  192. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), {}, name);
  193. float epsilon = 0.001;
  194. #if MEGDNN_AARCH64
  195. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_8X12X4_DOTPROD:96");
  196. #elif MEGDNN_ARMV7
  197. cb("IM2COLMATMUL:AARCH32_INT8_MK4_8X4X4_DOTPROD:96");
  198. #endif
  199. #undef cb
  200. }
  201. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_INT8x8x32_MK4_DOT) {
  202. UniformIntRNG rng{-50, 50};
  203. #define cb(name) \
  204. checker_conv_bias_common(get_nchw44_conv_bias_args({2, 3, 4, 5, 6, 7}, \
  205. ONLY_IDENTITY_NLMODE, \
  206. BR_AND_NO_BIASMODE, 1, \
  207. false, false, true), \
  208. handle(), &rng, epsilon, dtype::Int8(), \
  209. dtype::Int8(), dtype::Int32(), {}, name); \
  210. checker_conv_bias_common( \
  211. get_nchw44_conv_bias_args({1}, ONLY_IDENTITY_NLMODE, \
  212. BR_AND_BIAS_BIASMODE, 2, false, false, \
  213. true), \
  214. handle(), &rng, epsilon, dtype::Int8(), dtype::Int8(), \
  215. dtype::Int32(), {}, name);
  216. float epsilon = 0.001;
  217. #if MEGDNN_AARCH64
  218. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_8X12X4_DOTPROD:96");
  219. #elif MEGDNN_ARMV7
  220. cb("IM2COLMATMUL:AARCH32_INT8_MK4_8X4X4_DOTPROD:96");
  221. #endif
  222. #undef cb
  223. }
  224. #endif
  225. // clang-format on
  226. #if MEGDNN_AARCH64 || MEGDNN_ARMV7
  227. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_QUANTIZEDASYM) {
  228. NormalRNG rng(128.f);
  229. #define cb(name) \
  230. checker_conv_bias_common(get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, \
  231. false, false, true, true), \
  232. handle(), &rng, epsilon, \
  233. dtype::Quantized8Asymm(1.2f, (uint8_t)125), \
  234. dtype::Quantized8Asymm(1.3f, (uint8_t)129), \
  235. dtype::QuantizedS32(1.2 * 1.3), \
  236. dtype::Quantized8Asymm(50.3f, (uint8_t)120), \
  237. name); \
  238. checker_conv_bias_common( \
  239. get_conv_bias_args({1}, 2, false, false, false, true, true), \
  240. handle(), &rng, epsilon, \
  241. dtype::Quantized8Asymm(1.2f, (uint8_t)125), \
  242. dtype::Quantized8Asymm(1.3f, (uint8_t)129), \
  243. dtype::QuantizedS32(1.2 * 1.3), \
  244. dtype::Quantized8Asymm(50.3f, (uint8_t)120), name);
  245. float epsilon = 0.001;
  246. #if MEGDNN_AARCH64
  247. #if MGB_ENABLE_DOT
  248. cb("IM2COLMATMUL:AARCH64_QUINT8_K8X8X4_DOTPROD");
  249. #else
  250. cb("IM2COLMATMUL:AARCH64_QUINT8_K8X8X8");
  251. #endif
  252. #elif MEGDNN_ARMV7
  253. epsilon = 1;
  254. cb("IM2COLMATMUL:ARMV7_QUINT8_K4X8X8");
  255. #endif
  256. #undef cb
  257. }
  258. #endif
  259. #if MEGDNN_AARCH64 || MEGDNN_ARMV7
  260. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_QUINT8x8x32) {
  261. UniformIntRNG rng{-50, 50};
  262. float epsilon = 0.001;
  263. #define cb(name) \
  264. checker_conv_bias_common(get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, \
  265. false, true, true, false), \
  266. handle(), &rng, epsilon, \
  267. dtype::Quantized8Asymm(1.2f, (uint8_t)125), \
  268. dtype::Quantized8Asymm(1.3f, (uint8_t)129), \
  269. dtype::QuantizedS32(1.2 * 1.3), {}, name); \
  270. checker_conv_bias_common( \
  271. get_conv_bias_args({1}, 2, false, false, true, true, false), \
  272. handle(), &rng, epsilon, \
  273. dtype::Quantized8Asymm(1.2f, (uint8_t)125), \
  274. dtype::Quantized8Asymm(1.3f, (uint8_t)129), \
  275. dtype::QuantizedS32(1.2 * 1.3), {}, name);
  276. #if MEGDNN_AARCH64
  277. #if MGB_ENABLE_DOT
  278. cb("IM2COLMATMUL:AARCH64_QUINT8_K8X8X4_DOTPROD");
  279. #else
  280. cb("IM2COLMATMUL:AARCH64_QUINT8_K8X8X8");
  281. #endif
  282. #elif MEGDNN_ARMV7
  283. #if MGB_ENABLE_DOT
  284. cb("IM2COLMATMUL:AARCH32_QUINT8_K4X8X4");
  285. #endif
  286. cb("IM2COLMATMUL:ARMV7_QUINT8_K4X8X8");
  287. #endif
  288. #undef cb
  289. }
  290. TEST_F(ARM_COMMON_MULTI_THREADS, CONVBIAS_IM2COLMATMUL_INT8x8x16) {
  291. UniformIntRNG rng{-50, 50};
  292. float epsilon = 0.001;
  293. std::vector<conv_bias::TestArg> args_nchw44 =
  294. get_nchw44_conv_bias_args({2, 3, 4, 5, 6, 7}, ONLY_IDENTITY_NLMODE,
  295. BR_AND_BIAS_BIASMODE, 1, true);
  296. std::vector<conv_bias::TestArg> args_nchw44_1x1s2 =
  297. get_nchw44_conv_bias_args({1}, ONLY_IDENTITY_NLMODE,
  298. BR_AND_BIAS_BIASMODE, 2, true);
  299. #define cb(name) \
  300. checker_conv_bias_common( \
  301. get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, false, true), \
  302. handle(), &rng, epsilon, dtype::Int8{}, dtype::Int8{}, \
  303. dtype::Int16{}, dtype::Int16{}, name); \
  304. checker_conv_bias_common(get_conv_bias_args({1}, 2, false, false, true), \
  305. handle(), &rng, epsilon, dtype::Int8{}, \
  306. dtype::Int8{}, dtype::Int16{}, dtype::Int16{}, \
  307. name);
  308. #define cb_nchw44(name) \
  309. checker_conv_bias_common(args_nchw44, handle(), &rng, epsilon, \
  310. dtype::Int8{}, dtype::Int8{}, dtype::Int16{}, \
  311. dtype::Int16{}, name); \
  312. checker_conv_bias_common(args_nchw44_1x1s2, handle(), &rng, epsilon, \
  313. dtype::Int8{}, dtype::Int8{}, dtype::Int16{}, \
  314. dtype::Int16{}, name);
  315. #if MEGDNN_AARCH64
  316. cb("IM2COLMATMUL:AARCH64_INT8X8X16_K8X8X8");
  317. cb("IM2COLMATMUL:AARCH64_INT8X8X16_K4X4X16");
  318. cb_nchw44("IM2COLMATMUL:AARCH64_INT8X8X16_MK4_4X4X8");
  319. cb_nchw44("IM2COLMATMUL:AARCH64_INT8X8X16_MK4_16X12X4");
  320. #elif MEGDNN_ARMV7
  321. cb("IM2COLMATMUL:ARMV7_INT8X8X16_K4X8X8");
  322. cb("IM2COLMATMUL:ARMV7_INT8X8X16_K4X2X16");
  323. cb_nchw44("IM2COLMATMUL:ARMV7_INT8X8X16_MK4_K8X8X4");
  324. #endif
  325. cb("IM2COLMATMUL:ARM_COMMON_INT8X8X16");
  326. #undef cb
  327. #undef cb_nchw44
  328. }
  329. #endif
  330. #if __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
  331. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_FP16) {
  332. using namespace conv_bias;
  333. param::ConvBias cur_param;
  334. std::vector<conv_bias::TestArg> args =
  335. get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, false, false);
  336. std::vector<conv_bias::TestArg> args1 =
  337. get_conv_bias_args({1}, 2, false, false, false);
  338. args.insert(args.begin(), args1.begin(), args1.end());
  339. NormalRNG rng(1);
  340. #define cb(name) \
  341. checker_conv_bias_common(args, handle(), &rng, 0.03, dtype::Float16{}, \
  342. dtype::Float16{}, dtype::Float16{}, \
  343. dtype::Float16{}, name);
  344. #if MEGDNN_AARCH64
  345. cb("IM2COLMATMUL:AARCH64_F16_K8X24X1");
  346. #elif MEGDNN_ARMV7
  347. cb("IM2COLMATMUL:AARCH32_F16_K4X16X1");
  348. #endif
  349. #undef cb
  350. }
  351. #endif
  352. #if MEGDNN_AARCH64 || MEGDNN_ARMV7
  353. //! enable none dot algo now
  354. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_INT8x8x32NCHW44_S2) {
  355. using namespace conv_bias;
  356. std::vector<conv_bias::TestArg> args = get_nchw44_conv_bias_args(
  357. {2, 5, 7}, ONLY_IDENTITY_NLMODE, BR_AND_NO_BIASMODE, 2, false);
  358. #define cb(name) checker_conv_bias_mul_int8x8x32(args, handle(), name);
  359. #if MEGDNN_AARCH64
  360. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_4X4X16:96");
  361. #else
  362. cb("IM2COLMATMUL:ARMV7_INT8X8X32_MK4_4X2X16:96");
  363. #endif
  364. #undef cb
  365. }
  366. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_INT8x8x32NCHW44_S1) {
  367. using namespace conv_bias;
  368. std::vector<conv_bias::TestArg> args = get_nchw44_conv_bias_args(
  369. {3, 4, 6}, ONLY_IDENTITY_NLMODE, BR_AND_NO_BIASMODE, 1);
  370. #define cb(name) checker_conv_bias_mul_int8x8x32(args, handle(), name);
  371. #if MEGDNN_AARCH64
  372. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_4X4X16:96");
  373. #else
  374. cb("IM2COLMATMUL:ARMV7_INT8X8X32_MK4_4X2X16:96");
  375. #endif
  376. #undef cb
  377. }
  378. TEST_F(ARM_COMMON_MULTI_THREADS,
  379. CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_NCHW44_S2) {
  380. UniformIntRNG rng{-50, 50};
  381. #define cb(name) \
  382. checker_conv_bias_common( \
  383. get_nchw44_conv_bias_args({3, 4, 6}, QUAN_NLMODE, \
  384. BR_AND_NO_BIASMODE, 2), \
  385. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  386. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  387. dtype::QuantizedS8(60.25f), name);
  388. float epsilon = 0.001;
  389. #if MEGDNN_AARCH64
  390. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_4X4X16:96");
  391. #else
  392. epsilon = 1;
  393. cb("IM2COLMATMUL:ARMV7_INT8X8X32_MK4_4X2X16:96");
  394. #endif
  395. #undef cb
  396. }
  397. TEST_F(ARM_COMMON_MULTI_THREADS,
  398. CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_NCHW44_S1) {
  399. UniformIntRNG rng{-50, 50};
  400. #define cb(name) \
  401. checker_conv_bias_common( \
  402. get_nchw44_conv_bias_args({2, 5, 7}, QUAN_NLMODE, \
  403. BR_AND_NO_BIASMODE, 1), \
  404. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  405. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  406. dtype::QuantizedS8(60.25f), name);
  407. float epsilon = 0.001;
  408. #if MEGDNN_AARCH64
  409. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_4X4X16:96");
  410. #else
  411. epsilon = 1;
  412. cb("IM2COLMATMUL:ARMV7_INT8X8X32_MK4_4X2X16:96");
  413. #endif
  414. #undef cb
  415. }
  416. #if MEGDNN_AARCH64
  417. TEST_F(ARM_COMMON_MULTI_THREADS,
  418. CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_NCHW44_FUSE) {
  419. UniformIntRNG rng{-50, 50};
  420. #define cb(name) \
  421. checker_conv_bias_common( \
  422. get_nchw44_conv_bias_args({3}, QUAN_NLMODE, BR_AND_NO_BIASMODE, \
  423. 1), \
  424. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  425. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  426. dtype::QuantizedS8(60.25f), name);
  427. float epsilon = 0.001;
  428. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_4X4X16:96");
  429. #undef cb
  430. }
  431. #endif
  432. #endif
  433. #if MEGDNN_AARCH64
  434. #if MGB_ENABLE_DOT
  435. TEST_F(ARM_COMMON_MULTI_THREADS,
  436. CONV_BIAS_IM2COLMATMUL_QUANTIZEDSYM_NCHW44DOT_FUSE) {
  437. UniformIntRNG rng{-50, 50};
  438. #define cb(name) \
  439. checker_conv_bias_common( \
  440. get_nchw44_conv_bias_args({3}, QUAN_NLMODE, BR_AND_NO_BIASMODE, 1, \
  441. false, false, true), \
  442. handle(), &rng, epsilon, dtype::QuantizedS8(2.5f), \
  443. dtype::QuantizedS8(2.5f), dtype::QuantizedS32(6.25f), \
  444. dtype::QuantizedS8(60.25f), name);
  445. float epsilon = 0.001;
  446. cb("IM2COLMATMUL:AARCH64_INT8X8X32_MK4_8X12X4_DOTPROD:96");
  447. #undef cb
  448. }
  449. #endif
  450. #endif
  451. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COLMATMUL_INT8x8x32) {
  452. using namespace conv_bias;
  453. std::vector<conv_bias::TestArg> args =
  454. get_conv_bias_args({2, 3, 4, 5, 6, 7}, 1, false, true, true);
  455. std::vector<conv_bias::TestArg> args1 =
  456. get_conv_bias_args({1}, 2, false, true, true);
  457. args.insert(args.begin(), args1.begin(), args1.end());
  458. #define cb(name) checker_conv_bias_mul_int8x8x32(args, handle(), name);
  459. #if MEGDNN_AARCH64
  460. #if MGB_ENABLE_DOT
  461. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K8X12X4_DOTPROD");
  462. #else
  463. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K8X8X8");
  464. cb("IM2COLMATMUL:AARCH64_INT8X8X32_K4X4X16");
  465. #endif
  466. #elif MEGDNN_ARMV7
  467. #if MGB_ENABLE_DOT
  468. cb("IM2COLMATMUL:AARCH32_INT8_K6X8X4");
  469. #endif
  470. cb("IM2COLMATMUL:ARMV7_INT8X8X32_K4X8X8");
  471. #endif
  472. #if MEGDNN_ARMV7
  473. cb("IM2COLMATMUL:ARMV7_INT8X8X32_K4X2X16");
  474. #endif
  475. #undef cb
  476. }
  477. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COL_S1_MK4_PACK_F32) {
  478. using namespace conv_bias;
  479. std::vector<conv_bias::TestArg> args = get_nchw44_conv_bias_args(
  480. {2, 4, 7},FULL_NLMODE,BR_AND_BIAS_BIASMODE, 1);
  481. #if MEGDNN_AARCH64
  482. check_conv_bias(args, handle(), "IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  483. #elif MEGDNN_ARMV7
  484. check_conv_bias(args, handle(), "IM2COLMATMUL:ARMV7_F32_MK4_PACK_4X12");
  485. #endif
  486. }
  487. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COL_S2_MK4_PACK_F32) {
  488. using namespace conv_bias;
  489. std::vector<conv_bias::TestArg> args = get_nchw44_conv_bias_args(
  490. {3, 5, 6},FULL_NLMODE,BR_AND_BIAS_BIASMODE, 2);
  491. #define cb(name) check_conv_bias(args, handle(), name);
  492. #if MEGDNN_AARCH64
  493. cb("IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  494. #elif MEGDNN_ARMV7
  495. cb("IM2COLMATMUL:ARMV7_F32_MK4_PACK_4X12");
  496. #endif
  497. #undef cb
  498. }
  499. TEST_F(ARM_COMMON_MULTI_THREADS, CONV_BIAS_IM2COL_S2_MK4_PACK_F32_FUSE) {
  500. using namespace conv_bias;
  501. std::vector<conv_bias::TestArg> args = get_nchw44_conv_bias_args(
  502. {3},FULL_NLMODE,ALL_BIASMODE, 2);
  503. #define cb(name) check_conv_bias(args, handle(), name);
  504. #if MEGDNN_AARCH64
  505. cb("IM2COLMATMUL:AARCH64_F32_MK4_K8X12X1");
  506. #elif MEGDNN_ARMV7
  507. cb("IM2COLMATMUL:ARMV7_F32_MK4_PACK_4X12");
  508. #endif
  509. #undef cb
  510. }
  511. // vim: syntax=cpp.doxygen

MegEngine 安装包中集成了使用 GPU 运行代码所需的 CUDA 环境,不用区分 CPU 和 GPU 版。 如果想要运行 GPU 程序,请确保机器本身配有 GPU 硬件设备并安装好驱动。 如果你想体验在云端 GPU 算力平台进行深度学习开发的感觉,欢迎访问 MegStudio 平台