Megvii Engine Team
|
c2e9860feb
|
chore(license): remove all license in file header
GitOrigin-RevId: a0e31247a6
|
3 years ago |
Megvii Engine Team
|
47fe766310
|
feat(dnn/cuda): add implicit bmm kernels for large kernel depthwise convolution backward filter opr
GitOrigin-RevId: 932e7689e8
|
3 years ago |
Megvii Engine Team
|
369c2ccc5a
|
style(all): reformat c++ code
GitOrigin-RevId: 3ffd1b211f
|
3 years ago |
Megvii Engine Team
|
d69b59035d
|
feat(dnn): add an get_all_algorithms_safe interface
GitOrigin-RevId: e3734e4531
|
3 years ago |
Megvii Engine Team
|
0708bc780c
|
fix(dnn/cuda): disallow implicit dtype conversion in cublaslt matmul algos
disable tensor op matmul kernels when input and output tensors are in f32 data type to avoid potential accuracy loss
GitOrigin-RevId: 36859cba5a
|
3 years ago |
Megvii Engine Team
|
ff0e6be7b9
|
fix(dnn/cuda): fix cutlass tensorop kernels
do not compile cutlass tensorop kernels, when using cuda version less than 10.2
GitOrigin-RevId: d4c37d5f41
|
3 years ago |
Megvii Engine Team
|
336761253d
|
feat(dnn/cuda): add tensorcore matmul for fp16 data type
GitOrigin-RevId: 025c591f75
|
3 years ago |
Megvii Engine Team
|
ff755451d2
|
refactor(mgb): move algo's name from info to desc and delete some algo's unnecessary param() method
GitOrigin-RevId: 144ff547d1
|
4 years ago |
Megvii Engine Team
|
2de2222e46
|
feat(dnn/cuda): add cutlass batched gemv kernel for matmul operator
GitOrigin-RevId: 51702c4e79
|
4 years ago |
Megvii Engine Team
|
973d2a0ac2
|
feat(dnn/cuda): add cutlass matmul using split k parallel
GitOrigin-RevId: 650209e35f
|
4 years ago |
Megvii Engine Team
|
03c921f7c4
|
feat(dnn/cuda): add cutlass matmul impls
GitOrigin-RevId: 619c8c299c
|
4 years ago |