Megvii Engine Team
|
8fef78d06d
|
feat(dnn/cuda): add relayout format when width is an odd number
GitOrigin-RevId: f059f1f56d
|
4 years ago |
Megvii Engine Team
|
19a554d674
|
test(dnn/cuda): add testcase for transforming tensor layout between nchw and nchw64
GitOrigin-RevId: 75d579635a
|
4 years ago |
Megvii Engine Team
|
23032f50f2
|
feat(dnn/cuda): support float16 for index_incr_multi_axis_vec
GitOrigin-RevId: c2ae93d568
|
4 years ago |
Megvii Engine Team
|
938944027d
|
fix(mgb/dnn): fix cudnn8 convbias
GitOrigin-RevId: 0fdbfd258c
|
4 years ago |
Megvii Engine Team
|
1997b1a289
|
feat(dnn/cuda): add correlation kernel
GitOrigin-RevId: 25e58b61e6
|
4 years ago |
Megvii Engine Team
|
c3f8cf04fa
|
feat(dnn): add conv_bwd_data and conv_bwd_filter accuracy shake check
GitOrigin-RevId: 4069e083d2
|
4 years ago |
Megvii Engine Team
|
1e6ef3771f
|
feat(mgb/dnn): add accuracy shake checker
GitOrigin-RevId: 0bb52078a1
|
4 years ago |
Megvii Engine Team
|
ff755451d2
|
refactor(mgb): move algo's name from info to desc and delete some algo's unnecessary param() method
GitOrigin-RevId: 144ff547d1
|
4 years ago |
Megvii Engine Team
|
756c1eb7f2
|
fix(mgb/dnn): add cuda float naive matmul algo
GitOrigin-RevId: db7f7fc057
|
4 years ago |
Megvii Engine Team
|
68f2e59763
|
fix(mgb(ci)): fix tx1 ci testcase
GitOrigin-RevId: 7ac0c4b68d
|
4 years ago |
Megvii Engine Team
|
ba2ad46e54
|
feat(gopt): add deconv nchw4 int8 opt pass, add deconv nchw int8
GitOrigin-RevId: c0530a949e
|
4 years ago |
Megvii Engine Team
|
5d350fc843
|
feat(dnn/cuda): add deconv int8 and fix cutlass conv wrapper base on modify cutlass 2.4
GitOrigin-RevId: 49e0565e8a
|
4 years ago |
Megvii Engine Team
|
c82d88751a
|
fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility
GitOrigin-RevId: 771968f9ac
|
4 years ago |
Megvii Engine Team
|
97beae2fd8
|
fix(megdnn): fix megdnn benchmark testcase
GitOrigin-RevId: 726971474a
|
4 years ago |
Megvii Engine Team
|
2de2222e46
|
feat(dnn/cuda): add cutlass batched gemv kernel for matmul operator
GitOrigin-RevId: 51702c4e79
|
4 years ago |
Megvii Engine Team
|
973d2a0ac2
|
feat(dnn/cuda): add cutlass matmul using split k parallel
GitOrigin-RevId: 650209e35f
|
4 years ago |
Megvii Engine Team
|
03c921f7c4
|
feat(dnn/cuda): add cutlass matmul impls
GitOrigin-RevId: 619c8c299c
|
4 years ago |
Megvii Engine Team
|
cf27dd642c
|
fix(cuda): use cudnn8.0.4 as cu111 default libs
GitOrigin-RevId: 721ca73bae
|
4 years ago |
Megvii Engine Team
|
649e4dd750
|
test(cuda): fix test for cu111
GitOrigin-RevId: 04fe5eb23f
|
4 years ago |
Megvii Engine Team
|
c69359d00d
|
fix(dnn/cuda): disable cudnn conv_bias kernels for NCHW4_NCHW tensor format
due to illegal memory access errors caused by the kernels
GitOrigin-RevId: 584cedeeab
|
4 years ago |
Megvii Engine Team
|
0e3a6329ff
|
build(cuda): support cu111 build
GitOrigin-RevId: b3067ba4d5
|
4 years ago |
Megvii Engine Team
|
af42ce7e69
|
fix(megdnn): some fixes of execution policy
GitOrigin-RevId: 920f39bcb6
|
4 years ago |
Megvii Engine Team
|
821656aa4b
|
refactor(megdnn): refactor brute force algo in batched matmul
GitOrigin-RevId: 5c143ab3ac
|
4 years ago |
Megvii Engine Team
|
08ff62deb6
|
refactor(megdnn): refactor batched matmul algo in conv bias
GitOrigin-RevId: 64fda611ff
|
4 years ago |
Megvii Engine Team
|
8773926ef8
|
refactor(megdnn): refactor matmul algo in conv bias
GitOrigin-RevId: 932f7d6f81
|
4 years ago |
Megvii Engine Team
|
e4b71bdf64
|
refactor(megdnn): remove unnessary 1x1 algo
GitOrigin-RevId: 809f9b2bba
|
4 years ago |
Megvii Engine Team
|
b04ad06f84
|
refactor(megdnn): refactor matmul algo in conv backward filter
GitOrigin-RevId: bb65777583
|
4 years ago |
Megvii Engine Team
|
25089e520e
|
refactor(megdnn): refactor matmul algo in conv backward data
GitOrigin-RevId: 8de601df6e
|
4 years ago |
Megvii Engine Team
|
0d720653ac
|
refactor(megdnn): add default algo for convolution forward
GitOrigin-RevId: a12a7d399a
|
4 years ago |
Megvii Engine Team
|
659217acd2
|
refactor(megdnn): refactor bfloat16 convbias to recursive inteface
GitOrigin-RevId: 378194fb7f
|
4 years ago |
Megvii Engine Team
|
4a1d52c9c6
|
refactor(megdnn): refactor bfloat16 matmul to recursive inteface
GitOrigin-RevId: 641c508aec
|
4 years ago |
Megvii Engine Team
|
b8febaf91f
|
refactor(megdnn): refactor bfloat16 convolutionbackwardfilter to recursive inteface
GitOrigin-RevId: 37c08a5b8b
|
4 years ago |
Megvii Engine Team
|
f14e0c17e7
|
feat(mgb): add recursive for fastrun and megdnn test
GitOrigin-RevId: 743846f645
|
4 years ago |
Megvii Engine Team
|
364afec033
|
chore(mge): update copyright years
GitOrigin-RevId: 3c0690bcc1
|
4 years ago |
Megvii Engine Team
|
a85531dd0f
|
feat(mgb/opr): add tqt opr
GitOrigin-RevId: 49c62cd532
|
4 years ago |
Megvii Engine Team
|
c3a4b2225d
|
feat(dnn/cuda): add cutlass impls for fused convolution reformat operation
GitOrigin-RevId: 02ef559c3f
|
4 years ago |
Megvii Engine Team
|
5f44203d7b
|
feat(dnn/cuda): add a cutlass impl for fusing convolution and dimshuffle
GitOrigin-RevId: 3fc6faef01
|
4 years ago |
Megvii Engine Team
|
61f917fb8e
|
feat(dnn/cuda): add impl for fusing warp perspective and dimshuffle
GitOrigin-RevId: 51e025973f
|
4 years ago |
Megvii Engine Team
|
3bf73ff16f
|
feat(dnn): add cuda preprocess fusion
GitOrigin-RevId: d789c99e59
|
4 years ago |
Megvii Engine Team
|
142f31a875
|
perf(dnn/cuda): change conv_bias heu, prefer dnn chanwise impl, dislike dnn batch gemm conv1x1
GitOrigin-RevId: 323bf6073a
|
4 years ago |
Megvii Engine Team
|
a1877ee0fa
|
refactor(dnn): refactor algo interface, use algoinfo instead of global algorithm
GitOrigin-RevId: 479718ac75
|
4 years ago |
Megvii Engine Team
|
6f5d0febf1
|
perf(dnn/cuda): enhance performance for pooling forward
GitOrigin-RevId: 55fb2a9b25
|
4 years ago |
Megvii Engine Team
|
6856ce9ce2
|
feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format
GitOrigin-RevId: 29cd73f87b
|
4 years ago |
Megvii Engine Team
|
89ad33aeb3
|
feat(dnn/cuda): support weight preprocessing for cutlass algorithms
GitOrigin-RevId: 7b77579acd
|
4 years ago |
Megvii Engine Team
|
c03249c059
|
feat(dnn/opr): add megdnn fake quant opr
GitOrigin-RevId: 5a04b6da2f
|
4 years ago |
Megvii Engine Team
|
739f927c4c
|
feat(dnn/cuda): opt dp4a conv for small channel base on cutlass
GitOrigin-RevId: 2a74c35f27
|
4 years ago |
Megvii Engine Team
|
4aa277a203
|
refactor(dnn/cuda): misc
GitOrigin-RevId: 1f8f91a0cc
|
4 years ago |
Megvii Engine Team
|
ba66e1d039
|
feat(dnn): add nchw_fp32 nchw44_qint8 cuda dct
GitOrigin-RevId: 581e31fc20
|
4 years ago |
Megvii Engine Team
|
edb32495c6
|
feat(dnn/opr): add megdnn adaptive pooling opr
GitOrigin-RevId: 563ce65479
|
4 years ago |
Megvii Engine Team
|
310c805f20
|
fix(dnn/cuda): use kernel parameter instead of user constant memory
GitOrigin-RevId: 6080b24cc8
|
4 years ago |