Megvii Engine Team
|
5f44203d7b
|
feat(dnn/cuda): add a cutlass impl for fusing convolution and dimshuffle
GitOrigin-RevId: 3fc6faef01
|
4 years ago |
Megvii Engine Team
|
61f917fb8e
|
feat(dnn/cuda): add impl for fusing warp perspective and dimshuffle
GitOrigin-RevId: 51e025973f
|
4 years ago |
Megvii Engine Team
|
3bf73ff16f
|
feat(dnn): add cuda preprocess fusion
GitOrigin-RevId: d789c99e59
|
4 years ago |
Megvii Engine Team
|
142f31a875
|
perf(dnn/cuda): change conv_bias heu, prefer dnn chanwise impl, dislike dnn batch gemm conv1x1
GitOrigin-RevId: 323bf6073a
|
4 years ago |
Megvii Engine Team
|
a1877ee0fa
|
refactor(dnn): refactor algo interface, use algoinfo instead of global algorithm
GitOrigin-RevId: 479718ac75
|
4 years ago |
Megvii Engine Team
|
6f5d0febf1
|
perf(dnn/cuda): enhance performance for pooling forward
GitOrigin-RevId: 55fb2a9b25
|
4 years ago |
Megvii Engine Team
|
6856ce9ce2
|
feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format
GitOrigin-RevId: 29cd73f87b
|
4 years ago |
Megvii Engine Team
|
89ad33aeb3
|
feat(dnn/cuda): support weight preprocessing for cutlass algorithms
GitOrigin-RevId: 7b77579acd
|
4 years ago |
Megvii Engine Team
|
c03249c059
|
feat(dnn/opr): add megdnn fake quant opr
GitOrigin-RevId: 5a04b6da2f
|
4 years ago |
Megvii Engine Team
|
739f927c4c
|
feat(dnn/cuda): opt dp4a conv for small channel base on cutlass
GitOrigin-RevId: 2a74c35f27
|
4 years ago |
Megvii Engine Team
|
4aa277a203
|
refactor(dnn/cuda): misc
GitOrigin-RevId: 1f8f91a0cc
|
4 years ago |
Megvii Engine Team
|
ba66e1d039
|
feat(dnn): add nchw_fp32 nchw44_qint8 cuda dct
GitOrigin-RevId: 581e31fc20
|
4 years ago |
Megvii Engine Team
|
edb32495c6
|
feat(dnn/opr): add megdnn adaptive pooling opr
GitOrigin-RevId: 563ce65479
|
4 years ago |
Megvii Engine Team
|
310c805f20
|
fix(dnn/cuda): use kernel parameter instead of user constant memory
GitOrigin-RevId: 6080b24cc8
|
4 years ago |
Megvii Engine Team
|
3a03fa7a50
|
fix(dnn/cuda): disable pascal sass conv2d
GitOrigin-RevId: 385d066595
|
4 years ago |
Megvii Engine Team
|
a5fad7d07c
|
feat(dnn): add compile for riscv64
GitOrigin-RevId: fa0c163527
|
4 years ago |
Megvii Engine Team
|
76fa71573b
|
feat(dnn/cuda): add cutlass nchw4 convolution
GitOrigin-RevId: 93c9b212f4
|
4 years ago |
Megvii Engine Team
|
16324e3076
|
feat(dnn/cuda): add remap backward
GitOrigin-RevId: 1b1bcf5db3
|
5 years ago |
Megvii Engine Team
|
6e882c1a86
|
feat(whl/imperative): compat for build python whl imperative and legacy runtime
GitOrigin-RevId: 7f6629ae1f
|
4 years ago |
Megvii Engine Team
|
aeffcd5897
|
feat(dnn/cuda): integrate cutlass nchw32 tensorcore convolution
GitOrigin-RevId: 9d6c48ed99
|
4 years ago |
Megvii Engine Team
|
c7b6ef35c1
|
feat(dnn/cuda): add warp perspective backward mat idx
GitOrigin-RevId: b4b494bb69
|
5 years ago |
Megvii Engine Team
|
fc1ce273b7
|
fix(dnn/cuda): fix elemwise add cuda int8 bcast
GitOrigin-RevId: 568b60e8c9
|
4 years ago |
Megvii Engine Team
|
57bc36575f
|
style(dnn/cuda): format cuda elemwise code
GitOrigin-RevId: 246755ce20
|
4 years ago |
Megvii Engine Team
|
763b57add7
|
fix(dnn/cuda): fix INTMAX overflow in warp_perspective_cuda
GitOrigin-RevId: d7354e74e2
|
5 years ago |
Megvii Engine Team
|
7886ff9af0
|
feat(dnn): add relayout_format for nchw to nchw4 and ic <=4
GitOrigin-RevId: 07f2ee6c5b
|
5 years ago |
Megvii Engine Team
|
dedb7a3f14
|
feat(dnn/cuda): add cuda remap
GitOrigin-RevId: 40a2a2ce24
|
5 years ago |
Megvii Engine Team
|
69fe5ab3b3
|
feat(dnn/cuda): add conv2d-sass-kernel
GitOrigin-RevId: f284d5a4ce
|
5 years ago |
Megvii Engine Team
|
9f352b1c45
|
feat(megbrain/dnn): add indexing remap int32 for naive and cuda
GitOrigin-RevId: 5f66d51de4
|
5 years ago |
Megvii Engine Team
|
32c86211ee
|
fix(dnn/cuda): enable cuda algos for nchw quantized
GitOrigin-RevId: 4d1e167b86
|
5 years ago |
Megvii Engine Team
|
0b32056828
|
feat(dnn/naive): support weight preprocess interface in dnn
GitOrigin-RevId: 84791aacf9
|
5 years ago |
Megvii Engine Team
|
f354724220
|
fix(ci/megdnn_test/megbrain_test): split some
testcase, which caused by timeout
GitOrigin-RevId: ffed9d7820
|
5 years ago |
Megvii Engine Team
|
0293d58ade
|
feat(mge): add bfloat16 support
GitOrigin-RevId: a942ce6791
|
5 years ago |
Megvii Engine Team
|
1c4a64b2af
|
test(megdnn): skip fp16 test if compute capability less than 60
GitOrigin-RevId: 3377007b87
fix(dnn): fix include file redefined error
GitOrigin-RevId: 9d52f7ba50
|
5 years ago |
Megvii Engine Team
|
f582c1921e
|
fix(dnn/test): fix dnn assert and format code
GitOrigin-RevId: c684210c42
|
5 years ago |
luzzyzhang
|
16f052e916
|
fix(megdnn): change ver 60 to use cuda capability 50
fix(megdnn): if compute capability less than 60 skip fp16 test
|
5 years ago |
Megvii Engine Team
|
23478a0d53
|
test(dnn/cuda): fix cuda int8 test on sm60
GitOrigin-RevId: 66bab333e1
|
5 years ago |
Megvii Engine Team
|
5121626e98
|
refactor(dnn/parampack): remove parampacksplit
GitOrigin-RevId: ed00341d58
|
5 years ago |
Megvii Engine Team
|
b708f15d39
|
refactor(mgb/param_pack): use shared mem for param pack
GitOrigin-RevId: bc56f09037
|
5 years ago |
Megvii Engine Team
|
ca811c2c91
|
fix(mgb/opr/reduce): fix grad function of reduce mean and add testcase
GitOrigin-RevId: 390854bb2f
|
5 years ago |
Megvii Engine Team
|
f5833a5294
|
fix(dnn/cuda): fix cublas matmul on sm60
GitOrigin-RevId: 3fc0c30a23
|
5 years ago |
Megvii Engine Team
|
f91881ffdc
|
MegEngine: Initial commit of MegEngine.
GitOrigin-RevId: f0c8338beb
|
5 years ago |