Megvii Engine Team
|
afe9c4b50d
|
feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr
GitOrigin-RevId: 932e7689e8
|
3 years ago |
Megvii Engine Team
|
1da58ae17a
|
feat(dnn/cuda): add implicit bmm large kernel dwconv2d dgrad kernels
GitOrigin-RevId: fcb7974d62
|
3 years ago |
Megvii Engine Team
|
96050073a2
|
feat(dnn/cuda): add implicit bmm large kernel dwconv2d fprop impl
GitOrigin-RevId: feb09ebb58
|
3 years ago |
Megvii Engine Team
|
95ac055538
|
feat(dnn,mgb,imperative): add diag opr implement
GitOrigin-RevId: 43016ffa2b
|
3 years ago |
Megvii Engine Team
|
cbbca5fb10
|
feat(mge): add softmax op use cudnn api
GitOrigin-RevId: 7734ebf8c4
|
3 years ago |
Megvii Engine Team
|
20b42a8c3b
|
fix(dnn): add naive lstm kernel
GitOrigin-RevId: f08ef810cf
|
3 years ago |
Megvii Engine Team
|
2faa6ea5a9
|
Merge pull request #213 from kxz18:rnn
GitOrigin-RevId: 9e9215c115
|
3 years ago |
Megvii Engine Team
|
1999307015
|
feat(mgb/opr): add dropout kernel
GitOrigin-RevId: d248bd2005
|
3 years ago |
Megvii Engine Team
|
a93741815b
|
feat(mgb/opr): add layernorm forward and backward kernel
GitOrigin-RevId: 0cd484e753
|
3 years ago |
Megvii Engine Team
|
f6d9909460
|
feat(dnn): add elemwise multi type support i16xf32 and u8xf32
GitOrigin-RevId: 2fe469bb4e
|
3 years ago |
kxz@thumt102-1
|
8f48da7ffe
|
feat(mgb/opr): add cell level rnn/lstm and sequence level rnn/lstm
|
3 years ago |
Megvii Engine Team
|
2881934cb8
|
feat(dnn/check_non_finite): addmul scale to check_non_finite opr
GitOrigin-RevId: c35a219e52
|
3 years ago |
Megvii Engine Team
|
accb2d8d47
|
fix(mgb/serialize): fix flatbuffer compatibility issues
GitOrigin-RevId: e4771d6bc4
|
3 years ago |
Megvii Engine Team
|
b3e54eade1
|
feat(dnn/bn): use new cudnn BN kernel to support NHWC
GitOrigin-RevId: 9d80f2009d
|
3 years ago |
Megvii Engine Team
|
3977b7aa0b
|
feat(mgb/shuffle): add shuffle opr
GitOrigin-RevId: 80490a6f84
|
3 years ago |
Megvii Engine Team
|
4f28e14684
|
fix(dnn): fix compatibility broken of convolution format
GitOrigin-RevId: c17f02d03d
|
3 years ago |
Megvii Engine Team
|
fb49a2834f
|
refactor(mgb/dnn): refactor enum used in serializing
GitOrigin-RevId: e57af4a59c
|
3 years ago |
Megvii Engine Team
|
16678bb998
|
fix(dnn): fix_short_cutlass_name_gemm
GitOrigin-RevId: cc0a2db9da
|
3 years ago |
Megvii Engine Team
|
4c13bc7e1b
|
feat(dnn/cuda): add nhwc int8 deconv
GitOrigin-RevId: ad361a0f81
|
3 years ago |
Megvii Engine Team
|
11f022ff7c
|
feat(dnn/cuda): add nhwc int8 imma conv and conv fuse typecvt
GitOrigin-RevId: 229e1eb4be
|
3 years ago |
Megvii Engine Team
|
ff0e6be7b9
|
fix(dnn/cuda): fix cutlass tensorop kernels
do not compile cutlass tensorop kernels, when using cuda version less than 10.2
GitOrigin-RevId: d4c37d5f41
|
3 years ago |
Megvii Engine Team
|
336761253d
|
feat(dnn/cuda): add tensorcore matmul for fp16 data type
GitOrigin-RevId: 025c591f75
|
3 years ago |
Megvii Engine Team
|
2c4ee99227
|
fix(dnn): short cutlass filename in windows
GitOrigin-RevId: 83a43fdf87
|
3 years ago |
Megvii Engine Team
|
432592374d
|
build(dnn/cuda): fix cmake compile dependency for cutlass kernels
GitOrigin-RevId: ebe71f5a12
|
3 years ago |
Megvii Engine Team
|
eab6afab47
|
feat(mgb): add padding opr for megbrain
GitOrigin-RevId: 490e0c5d5a
|
4 years ago |
Megvii Engine Team
|
9b4b910dc1
|
feat(dnn/cuda): integrate cutlass operation table and replace all cutlass wrappers
GitOrigin-RevId: 2a70335441
|
3 years ago |
Megvii Engine Team
|
b18feaab33
|
feat(dnn/cuda): use cutlass remove shared load imma conv kernel
GitOrigin-RevId: 0b5574f526
|
4 years ago |
Megvii Engine Team
|
1af350c6d2
|
feat(dnn): add fill kernel
GitOrigin-RevId: d2cee3a7a0
|
3 years ago |
Megvii Engine Team
|
287cab49c2
|
fix(mgb/sereg): fix rng operator compatibility
GitOrigin-RevId: 66d1694035
|
3 years ago |
Megvii Engine Team
|
f76a2cc2c6
|
feat(mge/opr): add silu and gelu
GitOrigin-RevId: 75aa42947e
|
3 years ago |
Megvii Engine Team
|
f8b0f2cb91
|
build(dnn/cutlass): fix build for cutlass
GitOrigin-RevId: 9aa095fe84
|
3 years ago |
Megvii Engine Team
|
4eda338876
|
feat(dnn/cuda): generate cutlass kimpls using cmake and bazel
GitOrigin-RevId: da3bcfb85a
|
4 years ago |
Megvii Engine Team
|
894a2407c2
|
feat(dnn/cuda): add relayout format kernel for nchw <-> nhwc
GitOrigin-RevId: e11f3e5408
|
4 years ago |
Megvii Engine Team
|
5a14a89224
|
refactor(dnn/cuda): refactor cutlass kernel generator for gemm and gemv
GitOrigin-RevId: 11d78ab227
|
4 years ago |
Megvii Engine Team
|
4abf7bd36f
|
refactor(dnn/cuda): refactor kernel generator for cutlass convolution kernels
GitOrigin-RevId: 7882f9c68c
|
4 years ago |
Megvii Engine Team
|
66f70578c2
|
feat(dnn/cuda): add convolution with i8 input and i4 output
GitOrigin-RevId: 10512645d5
|
4 years ago |
Megvii Engine Team
|
43098fb8f1
|
feat(mge): add SlidingWindowTranspose opr
BREAKING CHANGE:
GitOrigin-RevId: 54d726d2fe
|
4 years ago |
Megvii Engine Team
|
b078dda90b
|
feat(mge/random): add some random op and remove random/distrbution.py
GitOrigin-RevId: 4c05ebc266
|
4 years ago |
Megvii Engine Team
|
f30c0e06a6
|
feat(mgb/opr): add lsq opr
GitOrigin-RevId: 45494a2b57
|
4 years ago |
Megvii Engine Team
|
12a0e61542
|
feat(dnn/cuda): add cuda elemwise int4
GitOrigin-RevId: 8a9aaec328
|
4 years ago |
Megvii Engine Team
|
71c2f61254
|
feat(dnn/cuda): add relayout format to support layout transform between NCHW and NCHW64
GitOrigin-RevId: 1445ecfabe
|
4 years ago |
Megvii Engine Team
|
ed92207585
|
feat(dnn/cuda): add conv bias impl for int4 data type using sass language
GitOrigin-RevId: ae3d3e1c98
|
4 years ago |
Megvii Engine Team
|
1525a02530
|
feat(mge/module): add python wrapper for unfold
GitOrigin-RevId: 562103186f
|
4 years ago |
Megvii Engine Team
|
1997b1a289
|
feat(dnn/cuda): add correlation kernel
GitOrigin-RevId: 25e58b61e6
|
4 years ago |
Megvii Engine Team
|
8494a1529e
|
chore(scripts): clarify and fix default value of bit combined enum
GitOrigin-RevId: 3716bf9bb5
|
4 years ago |
Megvii Engine Team
|
a3ea1f153c
|
feat(mgb/opr): add fast profile and combined Execution strategy
GitOrigin-RevId: 843dc3a790
|
4 years ago |
Megvii Engine Team
|
c82d88751a
|
fix(dnn/cuda): add cuda nchw int8 conv impl with nchw4 to fix cu111 compatibility
GitOrigin-RevId: 771968f9ac
|
4 years ago |
Megvii Engine Team
|
2de2222e46
|
feat(dnn/cuda): add cutlass batched gemv kernel for matmul operator
GitOrigin-RevId: 51702c4e79
|
4 years ago |
Megvii Engine Team
|
973d2a0ac2
|
feat(dnn/cuda): add cutlass matmul using split k parallel
GitOrigin-RevId: 650209e35f
|
4 years ago |
Megvii Engine Team
|
03c921f7c4
|
feat(dnn/cuda): add cutlass matmul impls
GitOrigin-RevId: 619c8c299c
|
4 years ago |