29 Commits (c42ce9370581debfad9ef0499deaa867b623cec0)

Author SHA1 Message Date
  Megvii Engine Team 8e5410e41f feat(cuda): add fp16 compute 16 kernel 3 years ago
  Megvii Engine Team e698ec20c2 feat(cuda): float16 depthwise large kernel conv compute fp32 3 years ago
  Megvii Engine Team 48406382ce feat(cuda): support float16 depthwise large kernel conv 3 years ago
  Megvii Engine Team 87a2aeebb1 perf(cuda): speedup chanwise conv with small feature map and large filter size 3 years ago
  Megvii Engine Team afe9c4b50d feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr 3 years ago
  Megvii Engine Team 38067472d2 fix(dnn/cuda): fix ci 3 years ago
  Megvii Engine Team ba2f0c2e48 fix(dnn/cuda): fix cudnn_conv algo of conv_bias opr for fp16 add z cases 3 years ago
  Megvii Engine Team 9b4cd92ba3 fix(mgb/dnn): fix cudnnConvBiasActivation crash on nchw32 int8 with oc > 256 3 years ago
  Megvii Engine Team 10af44abba fix(dnn/cuda): fix cudnn conv impl for nchw4_nchw hybrid layout 3 years ago
  Megvii Engine Team 369c2ccc5a style(all): reformat c++ code 3 years ago
  Megvii Engine Team 3b452d8c16 feat(mgb): cuda conv support nhwc format and fp16 dtype 3 years ago
  Megvii Engine Team 2aba0378b9 refactor(mgb/dnn): fix group conv is_available 3 years ago
  Megvii Engine Team 6ce212d2e0 refactor(mgb): refactor group conv 4 years ago
  Megvii Engine Team 7b2a76d1ee refactor(mgb): make conv handle noncontiguous tensors 4 years ago
  Megvii Engine Team 938944027d fix(mgb/dnn): fix cudnn8 convbias 4 years ago
  Megvii Engine Team c82d88751a fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility 4 years ago
  Megvii Engine Team cf27dd642c fix(cuda): use cudnn8.0.4 as cu111 default libs 4 years ago
  Megvii Engine Team 649e4dd750 test(cuda): fix test for cu111 4 years ago
  Megvii Engine Team c69359d00d fix(dnn/cuda): disable cudnn conv_bias kernels for NCHW4_NCHW tensor format 4 years ago
  Megvii Engine Team 08ff62deb6 refactor(megdnn): refactor batched matmul algo in conv bias 4 years ago
  Megvii Engine Team 8773926ef8 refactor(megdnn): refactor matmul algo in conv bias 4 years ago
  Megvii Engine Team e4b71bdf64 refactor(megdnn): remove unnessary 1x1 algo 4 years ago
  Megvii Engine Team 659217acd2 refactor(megdnn): refactor bfloat16 convbias to recursive inteface 4 years ago
  Megvii Engine Team 364afec033 chore(mge): update copyright years 4 years ago
  Megvii Engine Team 6856ce9ce2 feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format 4 years ago
  Megvii Engine Team 32c86211ee fix(dnn/cuda): enable cuda algos for nchw quantized 5 years ago
  Megvii Engine Team 23478a0d53 test(dnn/cuda): fix cuda int8 test on sm60 5 years ago
  Megvii Engine Team f5833a5294 fix(dnn/cuda): fix cublas matmul on sm60 5 years ago
  Megvii Engine Team f91881ffdc MegEngine: Initial commit of MegEngine. 5 years ago