GitOrigin-RevId: 563ce65479
release-1.1
@@ -683,6 +683,53 @@ protected: | |||
}; | |||
/** | |||
* \brief base class for AdaptivePooling | |||
*/ | |||
class AdaptivePoolingBase : public OperatorBase { | |||
DEF_OPR_IMPL_CTOR(AdaptivePoolingBase, OperatorBase); | |||
DEF_OPR_PARAM(AdaptivePooling); | |||
protected: | |||
param::Pooling deduce_pooling_param(const TensorLayout& src, | |||
const TensorLayout& dst); | |||
}; | |||
class AdaptivePoolingForward : public AdaptivePoolingBase { | |||
DEF_OPR_IMPL(AdaptivePoolingForward, AdaptivePoolingBase, 1, 1); | |||
public: | |||
/** | |||
* \param[in] src input tensor | |||
* \param[out] dst output tensor | |||
*/ | |||
virtual void exec(_megdnn_tensor_in src, _megdnn_tensor_out dst, | |||
_megdnn_workspace workspace) = 0; | |||
virtual size_t get_workspace_in_bytes(const TensorLayout& src, | |||
const TensorLayout& dst) = 0; | |||
}; | |||
using AdaptivePooling = AdaptivePoolingForward; | |||
class AdaptivePoolingBackward : public AdaptivePoolingBase { | |||
DEF_OPR_IMPL(AdaptivePoolingBackward, AdaptivePoolingBase, 3, 1); | |||
public: | |||
/** | |||
* \param[in] src the `src' parameter in AdaptivePoolingForward::exec | |||
* \param[in] dst the `dst' parameter in AdaptivePoolingForward::exec | |||
* \param[in] diff the backpropagated gradient wrt. dst | |||
* \param[out] grad the backpropagated gradient wrt. src | |||
*/ | |||
virtual void exec(_megdnn_tensor_in src, _megdnn_tensor_in dst, | |||
_megdnn_tensor_in diff, _megdnn_tensor_out grad, | |||
_megdnn_workspace workspace) = 0; | |||
virtual size_t get_workspace_in_bytes(const TensorLayout& src, | |||
const TensorLayout& dst, | |||
const TensorLayout& diff, | |||
const TensorLayout& grad) = 0; | |||
}; | |||
/** | |||
* \brief base class for Local | |||
*/ | |||
class LocalBase : public OperatorBase { | |||
@@ -0,0 +1,37 @@ | |||
/** | |||
* \file dnn/src/common/adaptive_pooling.cpp | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#include "megdnn/opr_param_defs.h" | |||
#include "megdnn/oprs.h" | |||
#include "src/common/utils.h" | |||
namespace megdnn { | |||
param::Pooling AdaptivePoolingBase::deduce_pooling_param( | |||
const TensorLayout& src, const TensorLayout& dst) { | |||
megdnn_assert(param().format == param::AdaptivePooling::Format::NCHW); | |||
size_t IH = src.shape[2], IW = src.shape[3], OH = dst.shape[2], | |||
OW = dst.shape[3]; | |||
param::Pooling ret; | |||
ret.mode = param().mode; | |||
ret.format = param().format; | |||
ret.pad_h = ret.pad_w = 0; | |||
ret.stride_h = floor(IH / OH); | |||
ret.stride_w = floor(IW / OW); | |||
ret.window_h = IH - (OH - 1) * ret.stride_h; | |||
ret.window_w = IW - (OW - 1) * ret.stride_w; | |||
return ret; | |||
} | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -199,6 +199,8 @@ private: | |||
cb(Remap) \ | |||
cb(RemapBackwardData) \ | |||
cb(RemapBackwardMat) \ | |||
cb(AdaptivePoolingForward) \ | |||
cb(AdaptivePoolingBackward) \ | |||
/*! | |||
* \brief specialize HandleImpl::create_operator for a single opr type; | |||
@@ -0,0 +1,53 @@ | |||
/** | |||
* \file dnn/src/cuda/adaptive_pooling/opr_impl.cpp | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#include "src/cuda/adaptive_pooling/opr_impl.h" | |||
#include "src/cuda/utils.h" | |||
namespace megdnn { | |||
namespace cuda { | |||
void AdaptivePoolingForwardImpl::exec(_megdnn_tensor_in src, | |||
_megdnn_tensor_out dst, | |||
_megdnn_workspace workspace) { | |||
auto opr = handle()->create_operator<PoolingForward>(); | |||
opr->param() = deduce_pooling_param(src.layout, dst.layout); | |||
opr->exec(src, dst, workspace); | |||
} | |||
size_t AdaptivePoolingForwardImpl::get_workspace_in_bytes( | |||
const TensorLayout& src, const TensorLayout& dst) { | |||
auto opr = handle()->create_operator<PoolingForward>(); | |||
opr->param() = deduce_pooling_param(src, dst); | |||
return opr->get_workspace_in_bytes(src, dst); | |||
} | |||
void AdaptivePoolingBackwardImpl::exec(_megdnn_tensor_in src, | |||
_megdnn_tensor_in dst, | |||
_megdnn_tensor_in diff, | |||
_megdnn_tensor_out grad, | |||
_megdnn_workspace workspace) { | |||
auto opr = handle()->create_operator<PoolingBackward>(); | |||
opr->param() = deduce_pooling_param(src.layout, dst.layout); | |||
opr->exec(src, dst, diff, grad, workspace); | |||
} | |||
size_t AdaptivePoolingBackwardImpl::get_workspace_in_bytes( | |||
const TensorLayout& src, const TensorLayout& dst, | |||
const TensorLayout& diff, const TensorLayout& grad) { | |||
auto opr = handle()->create_operator<PoolingBackward>(); | |||
opr->param() = deduce_pooling_param(src, dst); | |||
return opr->get_workspace_in_bytes(src, dst, diff, grad); | |||
} | |||
} // namespace cuda | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -0,0 +1,44 @@ | |||
/** | |||
* \file dnn/src/cuda/adaptive_pooling/opr_impl.h | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#pragma once | |||
#include "megdnn/oprs.h" | |||
#include "src/cuda/cudnn_wrapper.h" | |||
#include "src/cuda/utils.h" | |||
namespace megdnn { | |||
namespace cuda { | |||
class AdaptivePoolingForwardImpl final : public AdaptivePoolingForward { | |||
public: | |||
using AdaptivePoolingForward::AdaptivePoolingForward; | |||
void exec(_megdnn_tensor_in src, _megdnn_tensor_out dst, | |||
_megdnn_workspace workspace) override; | |||
size_t get_workspace_in_bytes(const TensorLayout& src, | |||
const TensorLayout& dst) override; | |||
}; | |||
class AdaptivePoolingBackwardImpl final : public AdaptivePoolingBackward { | |||
public: | |||
using AdaptivePoolingBackward::AdaptivePoolingBackward; | |||
void exec(_megdnn_tensor_in src, _megdnn_tensor_in dst, | |||
_megdnn_tensor_in diff, _megdnn_tensor_out grad, | |||
_megdnn_workspace workspace) override; | |||
size_t get_workspace_in_bytes(const TensorLayout& src, | |||
const TensorLayout& dst, | |||
const TensorLayout& diff, | |||
const TensorLayout& grad) override; | |||
}; | |||
} // namespace cuda | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -11,6 +11,7 @@ | |||
#include "src/common/handle_impl.h" | |||
#include "src/cuda/adaptive_pooling/opr_impl.h" | |||
#include "src/cuda/add_update/opr_impl.h" | |||
#include "src/cuda/argmxx/opr_impl.h" | |||
#include "src/cuda/argsort/opr_impl.h" | |||
@@ -0,0 +1,52 @@ | |||
/** | |||
* \file dnn/src/naive/adaptive_pooling/opr_impl.cpp | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#include "src/naive/adaptive_pooling/opr_impl.h" | |||
#include "src/common/opr_delegate.h" | |||
#include "src/common/utils.h" | |||
#include "src/naive/handle.h" | |||
namespace megdnn { | |||
namespace naive { | |||
void AdaptivePoolingForwardImpl::exec(_megdnn_tensor_in src, | |||
_megdnn_tensor_out dst, | |||
_megdnn_workspace workspace) { | |||
MEGDNN_DISPATCH_CPU_KERN(static_cast<naive::HandleImpl*>(handle()), { | |||
auto opr = inplace_cpu_handle()->create_operator<PoolingForward>(); | |||
opr->param() = deduce_pooling_param(src.layout, dst.layout); | |||
opr->exec(src, dst, workspace); | |||
}); | |||
} | |||
void AdaptivePoolingBackwardImpl::exec(_megdnn_tensor_in src, | |||
_megdnn_tensor_in dst, | |||
_megdnn_tensor_in diff, | |||
_megdnn_tensor_out grad, | |||
_megdnn_workspace workspace) { | |||
MEGDNN_DISPATCH_CPU_KERN(static_cast<naive::HandleImpl*>(handle()), { | |||
auto opr = inplace_cpu_handle()->create_operator<PoolingBackward>(); | |||
opr->param() = deduce_pooling_param(src.layout, dst.layout); | |||
opr->exec(src, dst, diff, grad, workspace); | |||
}); | |||
} | |||
size_t AdaptivePoolingBackwardImpl::get_workspace_in_bytes( | |||
const TensorLayout& src, const TensorLayout& dst, | |||
const TensorLayout& diff, const TensorLayout& grad) { | |||
auto opr = inplace_cpu_handle()->create_operator<PoolingBackward>(); | |||
opr->param() = deduce_pooling_param(src, dst); | |||
return opr->get_workspace_in_bytes(src, dst, diff, grad); | |||
} | |||
} // namespace naive | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -0,0 +1,43 @@ | |||
/** | |||
* \file dnn/src/naive/adaptive_pooling/opr_impl.h | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#pragma once | |||
#include "megdnn/oprs.h" | |||
#include "src/common/utils.h" | |||
namespace megdnn { | |||
namespace naive { | |||
class AdaptivePoolingForwardImpl : public AdaptivePoolingForward { | |||
public: | |||
using AdaptivePoolingForward::AdaptivePoolingForward; | |||
void exec(_megdnn_tensor_in src, _megdnn_tensor_out dst, | |||
_megdnn_workspace workspace) override; | |||
size_t get_workspace_in_bytes(const TensorLayout&, | |||
const TensorLayout&) override { | |||
return 0; | |||
} | |||
}; | |||
class AdaptivePoolingBackwardImpl : public AdaptivePoolingBackward { | |||
public: | |||
using AdaptivePoolingBackward::AdaptivePoolingBackward; | |||
void exec(_megdnn_tensor_in src, _megdnn_tensor_in dst, | |||
_megdnn_tensor_in diff, _megdnn_tensor_out grad, | |||
_megdnn_workspace workspace) override; | |||
size_t get_workspace_in_bytes(const TensorLayout& src, | |||
const TensorLayout& dst, | |||
const TensorLayout& diff, | |||
const TensorLayout& grad) override; | |||
}; | |||
} // namespace naive | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -13,6 +13,7 @@ | |||
#include "src/common/handle_impl.h" | |||
#include "src/naive/adaptive_pooling/opr_impl.h" | |||
#include "src/naive/add_update/opr_impl.h" | |||
#include "src/naive/argmxx/opr_impl.h" | |||
#include "src/naive/argsort/opr_impl.h" | |||
@@ -0,0 +1,55 @@ | |||
/** | |||
* \file dnn/test/common/adaptive_pooling.h | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#pragma once | |||
#include <cstddef> | |||
#include "megdnn/basic_types.h" | |||
#include "megdnn/opr_param_defs.h" | |||
namespace megdnn { | |||
namespace test { | |||
namespace adaptive_pooling { | |||
struct TestArg { | |||
param::AdaptivePooling param; | |||
TensorShape ishape; | |||
TensorShape oshape; | |||
TestArg(param::AdaptivePooling param, TensorShape ishape, | |||
TensorShape oshape) | |||
: param(param), ishape(ishape), oshape(oshape) {} | |||
}; | |||
inline std::vector<TestArg> get_args() { | |||
std::vector<TestArg> args; | |||
using Param = param::AdaptivePooling; | |||
using Mode = param::AdaptivePooling::Mode; | |||
for (size_t i = 36; i < 40; ++i) { | |||
args.emplace_back(Param{Mode::AVERAGE}, TensorShape{2, 3, i, i + 1}, | |||
TensorShape{2, 3, i - 4, i - 2}); | |||
args.emplace_back(Param{Mode::MAX}, TensorShape{2, 3, i, i + 1}, | |||
TensorShape{2, 3, i - 4, i - 2}); | |||
} | |||
for (size_t i = 5; i < 10; ++i) { | |||
args.emplace_back(Param{Mode::AVERAGE}, TensorShape{2, 3, i, i + 1}, | |||
TensorShape{2, 3, i - 3, i - 2}); | |||
args.emplace_back(Param{Mode::MAX}, TensorShape{2, 3, i, i + 1}, | |||
TensorShape{2, 3, i - 3, i - 2}); | |||
} | |||
return args; | |||
} | |||
} // namespace adaptive_pooling | |||
} // namespace test | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |
@@ -41,6 +41,8 @@ DEF(Images2NeibsForward, 2, true, true); | |||
DEF(Images2NeibsBackward, 2, true, false); | |||
DEF(PoolingForward, 2, true, true); | |||
DEF(PoolingBackward, 4, true, false); | |||
DEF(AdaptivePoolingForward, 2, true, false); | |||
DEF(AdaptivePoolingBackward, 4, true, false); | |||
DEF(LocalForward, 3, true, true); | |||
DEF(LocalBackwardData, 3, true, false); | |||
DEF(LocalBackwardFilter, 3, true, false); | |||
@@ -0,0 +1,97 @@ | |||
/** | |||
* \file dnn/test/cuda/adaptive_pooling.cpp | |||
* MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
* | |||
* Copyright (c) 2014-2020 Megvii Inc. All rights reserved. | |||
* | |||
* Unless required by applicable law or agreed to in writing, | |||
* software distributed under the License is distributed on an | |||
* "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
* implied. | |||
*/ | |||
#include "test/cuda/fixture.h" | |||
#include "megdnn/tensor_iter.h" | |||
#include "test/common/adaptive_pooling.h" | |||
#include "test/common/checker.h" | |||
#include "src/common/utils.h" | |||
#include "test/cuda/utils.h" | |||
#include <cudnn.h> | |||
#include "test/cuda/benchmark.h" | |||
namespace megdnn { | |||
namespace test { | |||
TEST_F(CUDA, ADAPTIVE_POOLING_FORWARD) { | |||
auto args = adaptive_pooling::get_args(); | |||
using Format = param::AdaptivePooling::Format; | |||
DType dtype = dtype::Float32(); | |||
for (auto&& arg : args) { | |||
auto param = arg.param; | |||
auto src = arg.ishape; | |||
auto dst = arg.oshape; | |||
param.format = Format::NCHW; | |||
Checker<AdaptivePooling> checker(handle_cuda()); | |||
checker.set_epsilon(1e-2); | |||
checker.set_param(param).set_dtype(0, dtype).set_dtype(1, dtype).exec( | |||
TensorShapeArray{src, dst, {}}); | |||
} | |||
} | |||
TEST_F(CUDA, ADAPTIVE_POOLING_BACKWARD) { | |||
auto args = adaptive_pooling::get_args(); | |||
for (auto&& arg : args) { | |||
Checker<AdaptivePoolingBackward> checker(handle_cuda()); | |||
TensorLayout ilayout = TensorLayout(arg.ishape, dtype::Float32()); | |||
TensorLayout olayout = TensorLayout(arg.oshape, dtype::Float32()); | |||
auto constraint = [this, | |||
arg](CheckerHelper::TensorValueArray& tensors_orig) { | |||
megdnn_assert(tensors_orig.size() == 4); | |||
auto opr = handle_cuda()->create_operator<AdaptivePoolingForward>(); | |||
opr->param() = arg.param; | |||
auto tensors_cuda_storage = CheckerHelper::alloc_tensors( | |||
handle_cuda(), | |||
{tensors_orig[0].layout, tensors_orig[1].layout}, 0); | |||
auto&& tensors_cuda = *tensors_cuda_storage; | |||
auto span = tensors_cuda[0].layout.span(); | |||
auto dst = static_cast<dt_byte*>(tensors_cuda[0].raw_ptr) + | |||
span.low_byte; | |||
auto src = static_cast<const dt_byte*>(tensors_orig[0].raw_ptr) + | |||
span.low_byte; | |||
megdnn_memcpy_H2D(handle_cuda(), dst, src, span.dist_byte()); | |||
auto workspace_size = opr->get_workspace_in_bytes( | |||
tensors_cuda[0].layout, tensors_cuda[1].layout); | |||
auto workspace_cuda = megdnn_malloc(handle_cuda(), workspace_size); | |||
Workspace workspace{static_cast<dt_byte*>(workspace_cuda), | |||
workspace_size}; | |||
opr->exec(tensors_cuda[0], tensors_cuda[1], workspace); | |||
megdnn_free(handle_cuda(), workspace_cuda); | |||
span = tensors_cuda[1].layout.span(); | |||
dst = static_cast<dt_byte*>(tensors_orig[1].raw_ptr) + | |||
span.low_byte; | |||
src = static_cast<const dt_byte*>(tensors_cuda[1].raw_ptr) + | |||
span.low_byte; | |||
megdnn_memcpy_D2H(handle_cuda(), dst, src, span.dist_byte()); | |||
}; | |||
DType dtype = dtype::Float32(); | |||
checker.set_tensors_constraint(constraint) | |||
.set_dtype(0, dtype) | |||
.set_dtype(1, dtype) | |||
.set_dtype(2, dtype) | |||
.set_dtype(3, dtype) | |||
.set_param(arg.param) | |||
.exec(TensorShapeArray{ilayout, olayout, olayout, ilayout}); | |||
} | |||
} | |||
} // namespace test | |||
} // namespace megdnn | |||
// vim: syntax=cpp.doxygen |