Megvii Engine Team
09dab38748
feat(cuda): support int1 simplewq conv
GitOrigin-RevId: 9c37c41bc7
3 years ago
Megvii Engine Team
87de704a46
feat(gopt): fuse conv h_swish
GitOrigin-RevId: a3d12991fb
3 years ago
Megvii Engine Team
d7b0994a3e
feat(cuda): add fp16 compute 16 kernel
GitOrigin-RevId: e03435be02
3 years ago
Megvii Engine Team
6b8a69d5b6
feat(cuda): float16 depthwise large kernel conv compute fp32
GitOrigin-RevId: 3050d48f26
3 years ago
Megvii Engine Team
bc385b5374
feat(cuda): support float16 depthwise large kernel conv
GitOrigin-RevId: fdc1b15fbc
3 years ago
Megvii Engine Team
72403e8929
perf(cuda): speedup chanwise conv with small feature map and large filter size
GitOrigin-RevId: e65b2ce856
3 years ago
Megvii Engine Team
47fe766310
feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr
GitOrigin-RevId: 932e7689e8
3 years ago
Megvii Engine Team
6cefabe734
fix(dnn/cuda): fix ci
GitOrigin-RevId: 8267e5f9dd
3 years ago
Megvii Engine Team
ba2f0c2e48
fix(dnn/cuda): fix cudnn_conv algo of conv_bias opr for fp16 add z cases
GitOrigin-RevId: b29b009de0
3 years ago
Megvii Engine Team
9b4cd92ba3
fix(mgb/dnn): fix cudnnConvBiasActivation crash on nchw32 int8 with oc > 256
GitOrigin-RevId: 20c0b90575
3 years ago
Megvii Engine Team
10af44abba
fix(dnn/cuda): fix cudnn conv impl for nchw4_nchw hybrid layout
the conv_bias algo *_IMPLICIT_GEMM in cudnn less than 8.0.0 is disabled due to the incorrect result for int8x4->f32 configs
GitOrigin-RevId: 7cc52d0a85
3 years ago
Megvii Engine Team
369c2ccc5a
style(all): reformat c++ code
GitOrigin-RevId: 3ffd1b211f
3 years ago
Megvii Engine Team
3b452d8c16
feat(mgb): cuda conv support nhwc format and fp16 dtype
GitOrigin-RevId: b8ddcd108a
3 years ago
Megvii Engine Team
2aba0378b9
refactor(mgb/dnn): fix group conv is_available
GitOrigin-RevId: b279909168
3 years ago
Megvii Engine Team
6ce212d2e0
refactor(mgb): refactor group conv
GitOrigin-RevId: 7afd312690
4 years ago
Megvii Engine Team
7b2a76d1ee
refactor(mgb): make conv handle noncontiguous tensors
GitOrigin-RevId: 86282709b3
4 years ago
Megvii Engine Team
938944027d
fix(mgb/dnn): fix cudnn8 convbias
GitOrigin-RevId: 0fdbfd258c
4 years ago
Megvii Engine Team
c82d88751a
fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility
GitOrigin-RevId: 771968f9ac
4 years ago
Megvii Engine Team
cf27dd642c
fix(cuda): use cudnn8.0.4 as cu111 default libs
GitOrigin-RevId: 721ca73bae
4 years ago
Megvii Engine Team
649e4dd750
test(cuda): fix test for cu111
GitOrigin-RevId: 04fe5eb23f
4 years ago
Megvii Engine Team
c69359d00d
fix(dnn/cuda): disable cudnn conv_bias kernels for NCHW4_NCHW tensor format
due to illegal memory access errors caused by the kernels
GitOrigin-RevId: 584cedeeab
4 years ago
Megvii Engine Team
08ff62deb6
refactor(megdnn): refactor batched matmul algo in conv bias
GitOrigin-RevId: 64fda611ff
4 years ago
Megvii Engine Team
8773926ef8
refactor(megdnn): refactor matmul algo in conv bias
GitOrigin-RevId: 932f7d6f81
4 years ago
Megvii Engine Team
e4b71bdf64
refactor(megdnn): remove unnessary 1x1 algo
GitOrigin-RevId: 809f9b2bba
4 years ago
Megvii Engine Team
659217acd2
refactor(megdnn): refactor bfloat16 convbias to recursive inteface
GitOrigin-RevId: 378194fb7f
4 years ago
Megvii Engine Team
364afec033
chore(mge): update copyright years
GitOrigin-RevId: 3c0690bcc1
4 years ago
Megvii Engine Team
6856ce9ce2
feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format
GitOrigin-RevId: 29cd73f87b
4 years ago
Megvii Engine Team
32c86211ee
fix(dnn/cuda): enable cuda algos for nchw quantized
GitOrigin-RevId: 4d1e167b86
5 years ago
Megvii Engine Team
23478a0d53
test(dnn/cuda): fix cuda int8 test on sm60
GitOrigin-RevId: 66bab333e1
5 years ago
Megvii Engine Team
f5833a5294
fix(dnn/cuda): fix cublas matmul on sm60
GitOrigin-RevId: 3fc0c30a23
5 years ago
Megvii Engine Team
f91881ffdc
MegEngine: Initial commit of MegEngine.
GitOrigin-RevId: f0c8338beb
5 years ago