Megvii Engine Team
87a2aeebb1
perf(cuda): speedup chanwise conv with small feature map and large filter size
GitOrigin-RevId: e65b2ce856
3 years ago
Megvii Engine Team
afe9c4b50d
feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr
GitOrigin-RevId: 932e7689e8
3 years ago
Megvii Engine Team
38067472d2
fix(dnn/cuda): fix ci
GitOrigin-RevId: 8267e5f9dd
3 years ago
Megvii Engine Team
ba2f0c2e48
fix(dnn/cuda): fix cudnn_conv algo of conv_bias opr for fp16 add z cases
GitOrigin-RevId: b29b009de0
3 years ago
Megvii Engine Team
9b4cd92ba3
fix(mgb/dnn): fix cudnnConvBiasActivation crash on nchw32 int8 with oc > 256
GitOrigin-RevId: 20c0b90575
3 years ago
Megvii Engine Team
10af44abba
fix(dnn/cuda): fix cudnn conv impl for nchw4_nchw hybrid layout
the conv_bias algo *_IMPLICIT_GEMM in cudnn less than 8.0.0 is disabled due to the incorrect result for int8x4->f32 configs
GitOrigin-RevId: 7cc52d0a85
3 years ago
Megvii Engine Team
369c2ccc5a
style(all): reformat c++ code
GitOrigin-RevId: 3ffd1b211f
3 years ago
Megvii Engine Team
3b452d8c16
feat(mgb): cuda conv support nhwc format and fp16 dtype
GitOrigin-RevId: b8ddcd108a
3 years ago
Megvii Engine Team
2aba0378b9
refactor(mgb/dnn): fix group conv is_available
GitOrigin-RevId: b279909168
3 years ago
Megvii Engine Team
6ce212d2e0
refactor(mgb): refactor group conv
GitOrigin-RevId: 7afd312690
4 years ago
Megvii Engine Team
7b2a76d1ee
refactor(mgb): make conv handle noncontiguous tensors
GitOrigin-RevId: 86282709b3
4 years ago
Megvii Engine Team
938944027d
fix(mgb/dnn): fix cudnn8 convbias
GitOrigin-RevId: 0fdbfd258c
4 years ago
Megvii Engine Team
c82d88751a
fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility
GitOrigin-RevId: 771968f9ac
4 years ago
Megvii Engine Team
cf27dd642c
fix(cuda): use cudnn8.0.4 as cu111 default libs
GitOrigin-RevId: 721ca73bae
4 years ago
Megvii Engine Team
649e4dd750
test(cuda): fix test for cu111
GitOrigin-RevId: 04fe5eb23f
4 years ago
Megvii Engine Team
c69359d00d
fix(dnn/cuda): disable cudnn conv_bias kernels for NCHW4_NCHW tensor format
due to illegal memory access errors caused by the kernels
GitOrigin-RevId: 584cedeeab
4 years ago
Megvii Engine Team
08ff62deb6
refactor(megdnn): refactor batched matmul algo in conv bias
GitOrigin-RevId: 64fda611ff
4 years ago
Megvii Engine Team
8773926ef8
refactor(megdnn): refactor matmul algo in conv bias
GitOrigin-RevId: 932f7d6f81
4 years ago
Megvii Engine Team
e4b71bdf64
refactor(megdnn): remove unnessary 1x1 algo
GitOrigin-RevId: 809f9b2bba
4 years ago
Megvii Engine Team
659217acd2
refactor(megdnn): refactor bfloat16 convbias to recursive inteface
GitOrigin-RevId: 378194fb7f
4 years ago
Megvii Engine Team
364afec033
chore(mge): update copyright years
GitOrigin-RevId: 3c0690bcc1
4 years ago
Megvii Engine Team
6856ce9ce2
feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format
GitOrigin-RevId: 29cd73f87b
4 years ago
Megvii Engine Team
32c86211ee
fix(dnn/cuda): enable cuda algos for nchw quantized
GitOrigin-RevId: 4d1e167b86
5 years ago
Megvii Engine Team
23478a0d53
test(dnn/cuda): fix cuda int8 test on sm60
GitOrigin-RevId: 66bab333e1
5 years ago
Megvii Engine Team
f5833a5294
fix(dnn/cuda): fix cublas matmul on sm60
GitOrigin-RevId: 3fc0c30a23
5 years ago
Megvii Engine Team
f91881ffdc
MegEngine: Initial commit of MegEngine.
GitOrigin-RevId: f0c8338beb
5 years ago