Megvii Engine Team
|
f6d9909460
|
feat(dnn): add elemwise multi type support i16xf32 and u8xf32
GitOrigin-RevId: 2fe469bb4e
|
3 years ago |
Megvii Engine Team
|
2881934cb8
|
feat(dnn/check_non_finite): addmul scale to check_non_finite opr
GitOrigin-RevId: c35a219e52
|
3 years ago |
Megvii Engine Team
|
accb2d8d47
|
fix(mgb/serialize): fix flatbuffer compatibility issues
GitOrigin-RevId: e4771d6bc4
|
3 years ago |
Megvii Engine Team
|
b3e54eade1
|
feat(dnn/bn): use new cudnn BN kernel to support NHWC
GitOrigin-RevId: 9d80f2009d
|
3 years ago |
Megvii Engine Team
|
3977b7aa0b
|
feat(mgb/shuffle): add shuffle opr
GitOrigin-RevId: 80490a6f84
|
3 years ago |
Megvii Engine Team
|
4f28e14684
|
fix(dnn): fix compatibility broken of convolution format
GitOrigin-RevId: c17f02d03d
|
3 years ago |
Megvii Engine Team
|
fb49a2834f
|
refactor(mgb/dnn): refactor enum used in serializing
GitOrigin-RevId: e57af4a59c
|
3 years ago |
Megvii Engine Team
|
16678bb998
|
fix(dnn): fix_short_cutlass_name_gemm
GitOrigin-RevId: cc0a2db9da
|
3 years ago |
Megvii Engine Team
|
4c13bc7e1b
|
feat(dnn/cuda): add nhwc int8 deconv
GitOrigin-RevId: ad361a0f81
|
3 years ago |
Megvii Engine Team
|
11f022ff7c
|
feat(dnn/cuda): add nhwc int8 imma conv and conv fuse typecvt
GitOrigin-RevId: 229e1eb4be
|
3 years ago |
Megvii Engine Team
|
ff0e6be7b9
|
fix(dnn/cuda): fix cutlass tensorop kernels
do not compile cutlass tensorop kernels, when using cuda version less than 10.2
GitOrigin-RevId: d4c37d5f41
|
3 years ago |
Megvii Engine Team
|
336761253d
|
feat(dnn/cuda): add tensorcore matmul for fp16 data type
GitOrigin-RevId: 025c591f75
|
3 years ago |
Megvii Engine Team
|
2c4ee99227
|
fix(dnn): short cutlass filename in windows
GitOrigin-RevId: 83a43fdf87
|
3 years ago |
Megvii Engine Team
|
432592374d
|
build(dnn/cuda): fix cmake compile dependency for cutlass kernels
GitOrigin-RevId: ebe71f5a12
|
3 years ago |
Megvii Engine Team
|
eab6afab47
|
feat(mgb): add padding opr for megbrain
GitOrigin-RevId: 490e0c5d5a
|
4 years ago |
Megvii Engine Team
|
9b4b910dc1
|
feat(dnn/cuda): integrate cutlass operation table and replace all cutlass wrappers
GitOrigin-RevId: 2a70335441
|
3 years ago |
Megvii Engine Team
|
b18feaab33
|
feat(dnn/cuda): use cutlass remove shared load imma conv kernel
GitOrigin-RevId: 0b5574f526
|
4 years ago |
Megvii Engine Team
|
1af350c6d2
|
feat(dnn): add fill kernel
GitOrigin-RevId: d2cee3a7a0
|
3 years ago |
Megvii Engine Team
|
287cab49c2
|
fix(mgb/sereg): fix rng operator compatibility
GitOrigin-RevId: 66d1694035
|
3 years ago |
Megvii Engine Team
|
f76a2cc2c6
|
feat(mge/opr): add silu and gelu
GitOrigin-RevId: 75aa42947e
|
3 years ago |
Megvii Engine Team
|
f8b0f2cb91
|
build(dnn/cutlass): fix build for cutlass
GitOrigin-RevId: 9aa095fe84
|
3 years ago |
Megvii Engine Team
|
4eda338876
|
feat(dnn/cuda): generate cutlass kimpls using cmake and bazel
GitOrigin-RevId: da3bcfb85a
|
4 years ago |
Megvii Engine Team
|
894a2407c2
|
feat(dnn/cuda): add relayout format kernel for nchw <-> nhwc
GitOrigin-RevId: e11f3e5408
|
4 years ago |
Megvii Engine Team
|
5a14a89224
|
refactor(dnn/cuda): refactor cutlass kernel generator for gemm and gemv
GitOrigin-RevId: 11d78ab227
|
4 years ago |
Megvii Engine Team
|
4abf7bd36f
|
refactor(dnn/cuda): refactor kernel generator for cutlass convolution kernels
GitOrigin-RevId: 7882f9c68c
|
4 years ago |
Megvii Engine Team
|
66f70578c2
|
feat(dnn/cuda): add convolution with i8 input and i4 output
GitOrigin-RevId: 10512645d5
|
4 years ago |
Megvii Engine Team
|
43098fb8f1
|
feat(mge): add SlidingWindowTranspose opr
BREAKING CHANGE:
GitOrigin-RevId: 54d726d2fe
|
4 years ago |
Megvii Engine Team
|
b078dda90b
|
feat(mge/random): add some random op and remove random/distrbution.py
GitOrigin-RevId: 4c05ebc266
|
4 years ago |
Megvii Engine Team
|
f30c0e06a6
|
feat(mgb/opr): add lsq opr
GitOrigin-RevId: 45494a2b57
|
4 years ago |
Megvii Engine Team
|
12a0e61542
|
feat(dnn/cuda): add cuda elemwise int4
GitOrigin-RevId: 8a9aaec328
|
4 years ago |
Megvii Engine Team
|
71c2f61254
|
feat(dnn/cuda): add relayout format to support layout transform between NCHW and NCHW64
GitOrigin-RevId: 1445ecfabe
|
4 years ago |
Megvii Engine Team
|
ed92207585
|
feat(dnn/cuda): add conv bias impl for int4 data type using sass language
GitOrigin-RevId: ae3d3e1c98
|
4 years ago |
Megvii Engine Team
|
1525a02530
|
feat(mge/module): add python wrapper for unfold
GitOrigin-RevId: 562103186f
|
4 years ago |
Megvii Engine Team
|
1997b1a289
|
feat(dnn/cuda): add correlation kernel
GitOrigin-RevId: 25e58b61e6
|
4 years ago |
Megvii Engine Team
|
8494a1529e
|
chore(scripts): clarify and fix default value of bit combined enum
GitOrigin-RevId: 3716bf9bb5
|
4 years ago |
Megvii Engine Team
|
a3ea1f153c
|
feat(mgb/opr): add fast profile and combined Execution strategy
GitOrigin-RevId: 843dc3a790
|
4 years ago |
Megvii Engine Team
|
c82d88751a
|
fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility
GitOrigin-RevId: 771968f9ac
|
4 years ago |
Megvii Engine Team
|
2de2222e46
|
feat(dnn/cuda): add cutlass batched gemv kernel for matmul operator
GitOrigin-RevId: 51702c4e79
|
4 years ago |
Megvii Engine Team
|
973d2a0ac2
|
feat(dnn/cuda): add cutlass matmul using split k parallel
GitOrigin-RevId: 650209e35f
|
4 years ago |
Megvii Engine Team
|
03c921f7c4
|
feat(dnn/cuda): add cutlass matmul impls
GitOrigin-RevId: 619c8c299c
|
4 years ago |
Megvii Engine Team
|
ad87f78a14
|
chore(imperative): refine tblgen for generating op name
GitOrigin-RevId: f47ceae726
|
4 years ago |
Megvii Engine Team
|
55042195d4
|
chore(winograd): add Convolutionv2 param
GitOrigin-RevId: 1a9e2ea340
|
4 years ago |
Megvii Engine Team
|
a85531dd0f
|
feat(mgb/opr): add tqt opr
GitOrigin-RevId: 49c62cd532
|
4 years ago |
Megvii Engine Team
|
61f917fb8e
|
feat(dnn/cuda): add impl for fusing warp perspective and dimshuffle
GitOrigin-RevId: 51e025973f
|
4 years ago |
Megvii Engine Team
|
fc0fcd2f7f
|
chore(winograd): remove winograd transform code
GitOrigin-RevId: 78c3cfceae
|
4 years ago |
Megvii Engine Team
|
69e3e32240
|
feat(imperative): auto generated opdef header and python binding
GitOrigin-RevId: d2f22ad5fe
|
4 years ago |
Megvii Engine Team
|
3bf73ff16f
|
feat(dnn): add cuda preprocess fusion
GitOrigin-RevId: d789c99e59
|
4 years ago |
Megvii Engine Team
|
6856ce9ce2
|
feat(dnn): support conv bias activation for nchw4 input tensor format and nchw output tensor format
GitOrigin-RevId: 29cd73f87b
|
4 years ago |
Megvii Engine Team
|
c03249c059
|
feat(dnn/opr): add megdnn fake quant opr
GitOrigin-RevId: 5a04b6da2f
|
4 years ago |
Megvii Engine Team
|
ba66e1d039
|
feat(dnn): add nchw_fp32 nchw44_qint8 cuda dct
GitOrigin-RevId: 581e31fc20
|
4 years ago |