Megvii Engine Team
8e5410e41f
feat(cuda): add fp16 compute 16 kernel
GitOrigin-RevId: e03435be02
3 years ago
Megvii Engine Team
472e2f9655
refactor(cuda): depthwish large kernel
GitOrigin-RevId: dade8710b4
3 years ago
Megvii Engine Team
e698ec20c2
feat(cuda): float16 depthwise large kernel conv compute fp32
GitOrigin-RevId: 3050d48f26
3 years ago
Megvii Engine Team
48406382ce
feat(cuda): support float16 depthwise large kernel conv
GitOrigin-RevId: fdc1b15fbc
3 years ago
Megvii Engine Team
7042f76b34
perf(cuda): speedup conv backward data with small feature map and large filter size
GitOrigin-RevId: 85592bca6b
3 years ago
Megvii Engine Team
87a2aeebb1
perf(cuda): speedup chanwise conv with small feature map and large filter size
GitOrigin-RevId: e65b2ce856
3 years ago
Megvii Engine Team
2293385e93
feat(mge): add conv padding mode
GitOrigin-RevId: 147ced856e
3 years ago
Megvii Engine Team
afe9c4b50d
feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr
GitOrigin-RevId: 932e7689e8
3 years ago
Megvii Engine Team
38067472d2
fix(dnn/cuda): fix ci
GitOrigin-RevId: 8267e5f9dd
3 years ago
Megvii Engine Team
1da58ae17a
feat(dnn/cuda): add implicit bmm large kernel dwconv2d dgrad kernels
GitOrigin-RevId: fcb7974d62
3 years ago
Megvii Engine Team
96050073a2
feat(dnn/cuda): add implicit bmm large kernel dwconv2d fprop impl
GitOrigin-RevId: feb09ebb58
3 years ago
Megvii Engine Team
95ac055538
feat(dnn,mgb,imperative): add diag opr implement
GitOrigin-RevId: 43016ffa2b
3 years ago
Megvii Engine Team
cbbca5fb10
feat(mge): add softmax op use cudnn api
GitOrigin-RevId: 7734ebf8c4
3 years ago
Megvii Engine Team
82be0aaced
test(dnn): fix compute capability requirement for NCHWX test
GitOrigin-RevId: d2f8022be1
3 years ago
Megvii Engine Team
1999307015
feat(mgb/opr): add dropout kernel
GitOrigin-RevId: d248bd2005
3 years ago
Megvii Engine Team
a93741815b
feat(mgb/opr): add layernorm forward and backward kernel
GitOrigin-RevId: 0cd484e753
3 years ago
Megvii Engine Team
2696e4efaa
feat(dnn): add float16 for remap backward
GitOrigin-RevId: 0263030051
3 years ago
Megvii Engine Team
11d75fecb5
feat(dnn/check_non_finite): add batch check_non_finite
GitOrigin-RevId: e108133282
3 years ago
Megvii Engine Team
ba2f0c2e48
fix(dnn/cuda): fix cudnn_conv algo of conv_bias opr for fp16 add z cases
GitOrigin-RevId: b29b009de0
3 years ago
Megvii Engine Team
c85631aa77
feat(dnn): use ref ptr interface for all backends
GitOrigin-RevId: f65feae5cc
3 years ago
Megvii Engine Team
89186edc5d
fix(dnn): correct reduce/argmxx/fakequant calculation with nan
GitOrigin-RevId: 7e78bdae91
3 years ago
Megvii Engine Team
68cdabd288
feat(opr): indexing_multi_axis_vec support nd index
GitOrigin-RevId: 07b1248bdc
3 years ago
Megvii Engine Team
9b4cd92ba3
fix(mgb/dnn): fix cudnnConvBiasActivation crash on nchw32 int8 with oc > 256
GitOrigin-RevId: 20c0b90575
3 years ago
Megvii Engine Team
10af44abba
fix(dnn/cuda): fix cudnn conv impl for nchw4_nchw hybrid layout
the conv_bias algo *_IMPLICIT_GEMM in cudnn less than 8.0.0 is disabled due to the incorrect result for int8x4->f32 configs
GitOrigin-RevId: 7cc52d0a85
3 years ago
Megvii Engine Team
5885b137fa
feat(dnn/arm): support layout like NHWC channel like broadcast on arm
GitOrigin-RevId: fb4300004c
3 years ago
Megvii Engine Team
369c2ccc5a
style(all): reformat c++ code
GitOrigin-RevId: 3ffd1b211f
3 years ago
Megvii Engine Team
f5cb21ed3a
fix(mgb/opr): add non finite check
GitOrigin-RevId: a9fcd0a350
3 years ago
Megvii Engine Team
bde5cf3564
feat(dnn): add resize linear for arm
GitOrigin-RevId: 14ac5bda3f
3 years ago
Megvii Engine Team
3d3666b6e0
test(dnn/bn): add compatible configs for NHWC BN
GitOrigin-RevId: ac757ca307
3 years ago
Megvii Engine Team
3977b7aa0b
feat(mgb/shuffle): add shuffle opr
GitOrigin-RevId: 80490a6f84
3 years ago
Megvii Engine Team
17371e79b9
fix(dnn/reduce): fix reduce_mean o16c32 is incorrect for large tensor
GitOrigin-RevId: ebf03d814a
3 years ago
Megvii Engine Team
8b40f57738
feat(mgb/dnn): add conv1x1 algo for matrix mul
GitOrigin-RevId: 585b2c045a
3 years ago
Megvii Engine Team
d69b59035d
feat(dnn): add an get_all_algorithms_safe interface
GitOrigin-RevId: e3734e4531
3 years ago
Megvii Engine Team
8b94f49328
fix(dnn/cuda): fix elemwise and relayout int4 bug when last shape is 1
GitOrigin-RevId: e7d64c4987
3 years ago
Megvii Engine Team
722aecd437
feat(mgb): support fp16 nhwc backward
GitOrigin-RevId: 954ac6405a
3 years ago
Megvii Engine Team
0708bc780c
fix(dnn/cuda): disallow implicit dtype conversion in cublaslt matmul algos
disable tensor op matmul kernels when input and output tensors are in f32 data type to avoid potential accuracy loss
GitOrigin-RevId: 36859cba5a
3 years ago
Megvii Engine Team
4c13bc7e1b
feat(dnn/cuda): add nhwc int8 deconv
GitOrigin-RevId: ad361a0f81
3 years ago
Megvii Engine Team
11f022ff7c
feat(dnn/cuda): add nhwc int8 imma conv and conv fuse typecvt
GitOrigin-RevId: 229e1eb4be
3 years ago
Megvii Engine Team
67575d582c
feat(mge/opr): add interpolate bilinear mode
GitOrigin-RevId: f7023a3fd3
3 years ago
Megvii Engine Team
0558b2123d
feat(mge/opr): add interpolate nearest mode
GitOrigin-RevId: d384b87f50
3 years ago
Megvii Engine Team
c25125e3d2
perf(dnn/cuda): sass int8 epilogue remove shared load
GitOrigin-RevId: 2b49f5069b
3 years ago
Megvii Engine Team
ff0e6be7b9
fix(dnn/cuda): fix cutlass tensorop kernels
do not compile cutlass tensorop kernels, when using cuda version less than 10.2
GitOrigin-RevId: d4c37d5f41
3 years ago
Megvii Engine Team
336761253d
feat(dnn/cuda): add tensorcore matmul for fp16 data type
GitOrigin-RevId: 025c591f75
3 years ago
Megvii Engine Team
eab6afab47
feat(mgb): add padding opr for megbrain
GitOrigin-RevId: 490e0c5d5a
4 years ago
Megvii Engine Team
b18feaab33
feat(dnn/cuda): use cutlass remove shared load imma conv kernel
GitOrigin-RevId: 0b5574f526
4 years ago
Megvii Engine Team
1af350c6d2
feat(dnn): add fill kernel
GitOrigin-RevId: d2cee3a7a0
3 years ago
Megvii Engine Team
3eb0505f9b
feat(imperative): add support for quantized conv transpose2d
GitOrigin-RevId: ffd6431299
3 years ago
Megvii Engine Team
3b452d8c16
feat(mgb): cuda conv support nhwc format and fp16 dtype
GitOrigin-RevId: b8ddcd108a
3 years ago
Megvii Engine Team
2aba0378b9
refactor(mgb/dnn): fix group conv is_available
GitOrigin-RevId: b279909168
3 years ago
Megvii Engine Team
4a92346b7a
refactor(mgb): refactor group conv3d
GitOrigin-RevId: 15360a3a41
3 years ago