You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by GitBox <gi...@apache.org> on 2018/01/10 22:25:07 UTC
[GitHub] cjolivier01 closed pull request #9371: Fix unit test build
cjolivier01 closed pull request #9371: Fix unit test build
URL: https://github.com/apache/incubator-mxnet/pull/9371
This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:
As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):
diff --git a/tests/cpp/misc/memory_test.cc b/tests/cpp/misc/memory_test.cc
index 8f4e8c25e8..7c83e461bd 100644
--- a/tests/cpp/misc/memory_test.cc
+++ b/tests/cpp/misc/memory_test.cc
@@ -64,40 +64,40 @@ TEST(MEMORY_TEST, MemsetAndMemcopyPerformance) {
std::cout << "====================================" << std::endl
<< "Data size: " << test::pretty_num(test_size) << std::endl << std::flush;
- std::unique_ptr<uint8_t> buffer_1(new uint8_t[test_size]), buffer_2(new uint8_t[test_size]);
- uint8_t *src = buffer_1.get(), *dest = buffer_2.get();
+ std::unique_ptr<float> buffer_1(new float[test_size]), buffer_2(new float[test_size]);
+ float *src = buffer_1.get(), *dest = buffer_2.get();
for (size_t x = 0; x < 5; ++x) {
// Init memory with different values
- memset(src, 3, test_size);
- memset(dest, 255, test_size); // wipe out some/all of src cache
+ memset(src, 3, test_size * sizeof(float));
+ memset(dest, 255, test_size * sizeof(float)); // wipe out some/all of src cache
// memset
- uint64_t start = test::perf::getNannoTickCount();
- memset(src, 123, test_size);
- const uint64_t memset_time = test::perf::getNannoTickCount() - start;
+ uint64_t start = mxnet::test::perf::getNannoTickCount();
+ memset(src, 0, test_size * sizeof(float));
+ const uint64_t memset_time = mxnet::test::perf::getNannoTickCount() - start;
- start = test::perf::getNannoTickCount();
+ start = mxnet::test::perf::getNannoTickCount();
#pragma omp parallel for num_threads(GetOMPThreadCount())
for (int i = 0; i < static_cast<int>(test_size); ++i) {
- src[i] = 42;
+ src[i] = 42.0f;
}
- const uint64_t omp_set_time = test::perf::getNannoTickCount() - start;
+ const uint64_t omp_set_time = mxnet::test::perf::getNannoTickCount() - start;
- start = test::perf::getNannoTickCount();
- memcpy(dest, src, test_size);
- const uint64_t memcpy_time = test::perf::getNannoTickCount() - start;
+ start = mxnet::test::perf::getNannoTickCount();
+ memcpy(dest, src, test_size * sizeof(float));
+ const uint64_t memcpy_time = mxnet::test::perf::getNannoTickCount() - start;
// bounce the cache and dirty logic
- memset(src, 6, test_size);
- memset(dest, 200, test_size);
+ memset(src, 6, test_size * sizeof(float));
+ memset(dest, 200, test_size * sizeof(float));
- start = test::perf::getNannoTickCount();
+ start = mxnet::test::perf::getNannoTickCount();
#pragma omp parallel for num_threads(GetOMPThreadCount())
for (int i = 0; i < static_cast<int>(test_size); ++i) {
dest[i] = src[i];
}
- const uint64_t omp_copy_time = test::perf::getNannoTickCount() - start;
+ const uint64_t omp_copy_time = mxnet::test::perf::getNannoTickCount() - start;
memset_times.push_back(memset_time);
omp_set_times.push_back(omp_set_time);
diff --git a/tests/cpp/operator/batchnorm_test.cc b/tests/cpp/operator/batchnorm_test.cc
index d9b426a815..179e42a383 100644
--- a/tests/cpp/operator/batchnorm_test.cc
+++ b/tests/cpp/operator/batchnorm_test.cc
@@ -392,17 +392,17 @@ class BatchNormValidator : public test::op::Validator<DType, AccReal> {
// Input
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInput,
- op::batchnorm::kData));
+ mxnet::op::batchnorm::kData));
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInput,
- op::batchnorm::kGamma));
+ mxnet::op::batchnorm::kGamma));
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInput,
- op::batchnorm::kBeta));
+ mxnet::op::batchnorm::kBeta));
// Output
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kOutput,
- op::batchnorm::kOut));
+ mxnet::op::batchnorm::kOut));
CHECK_EQ(info_2.prop_->getParam().use_global_stats,
info_1.prop_->getParam().use_global_stats);
@@ -410,29 +410,29 @@ class BatchNormValidator : public test::op::Validator<DType, AccReal> {
// Aux
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kAux,
- op::batchnorm::kMovingMean));
+ mxnet::op::batchnorm::kMovingMean));
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kAux,
- op::batchnorm::kMovingVar));
+ mxnet::op::batchnorm::kMovingVar));
#endif
if (!info_2.prop_->getParam().use_global_stats) {
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kOutput,
- op::batchnorm::kMean));
+ mxnet::op::batchnorm::kMean));
// InGrad
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInGrad,
- op::batchnorm::kData));
+ mxnet::op::batchnorm::kData));
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInGrad,
- op::batchnorm::kGamma));
+ mxnet::op::batchnorm::kGamma));
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInGrad,
- op::batchnorm::kBeta));
+ mxnet::op::batchnorm::kBeta));
// OutGrad
EXPECT_TRUE(compare(*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kOutGrad,
- op::batchnorm::kData));
+ mxnet::op::batchnorm::kData));
}
}
};
@@ -488,16 +488,16 @@ static StreamType& dumpF(StreamType *os,
*os << "=============================" << std::endl;
}
typedef typename OperatorExecutor::BlobVectorType BlobVectorType;
- PRT(os, *prop.executor_, BlobVectorType::kInput, op::batchnorm::kData);
- PRT(os, *prop.executor_, BlobVectorType::kInput, op::batchnorm::kGamma);
- PRT(os, *prop.executor_, BlobVectorType::kInput, op::batchnorm::kBeta);
+ PRT(os, *prop.executor_, BlobVectorType::kInput, mxnet::op::batchnorm::kData);
+ PRT(os, *prop.executor_, BlobVectorType::kInput, mxnet::op::batchnorm::kGamma);
+ PRT(os, *prop.executor_, BlobVectorType::kInput, mxnet::op::batchnorm::kBeta);
- PRT(os, *prop.executor_, BlobVectorType::kAux, op::batchnorm::kMovingMean);
- PRT(os, *prop.executor_, BlobVectorType::kAux, op::batchnorm::kMovingVar);
+ PRT(os, *prop.executor_, BlobVectorType::kAux, mxnet::op::batchnorm::kMovingMean);
+ PRT(os, *prop.executor_, BlobVectorType::kAux, mxnet::op::batchnorm::kMovingVar);
- PRT(os, *prop.executor_, BlobVectorType::kOutput, op::batchnorm::kOut);
- PRT(os, *prop.executor_, BlobVectorType::kOutput, op::batchnorm::kMean);
- PRT(os, *prop.executor_, BlobVectorType::kOutput, op::batchnorm::kVar);
+ PRT(os, *prop.executor_, BlobVectorType::kOutput, mxnet::op::batchnorm::kOut);
+ PRT(os, *prop.executor_, BlobVectorType::kOutput, mxnet::op::batchnorm::kMean);
+ PRT(os, *prop.executor_, BlobVectorType::kOutput, mxnet::op::batchnorm::kVar);
}
return *os;
}
@@ -515,14 +515,14 @@ static StreamType& dumpB(StreamType *os,
}
typedef typename OperatorExecutor::BlobVectorType BlobVectorType;
- PRT(os, *prop.executor_, BlobVectorType::kInGrad, op::batchnorm::kData);
- PRT(os, *prop.executor_, BlobVectorType::kInGrad, op::batchnorm::kGamma);
- PRT(os, *prop.executor_, BlobVectorType::kInGrad, op::batchnorm::kBeta);
+ PRT(os, *prop.executor_, BlobVectorType::kInGrad, mxnet::op::batchnorm::kData);
+ PRT(os, *prop.executor_, BlobVectorType::kInGrad, mxnet::op::batchnorm::kGamma);
+ PRT(os, *prop.executor_, BlobVectorType::kInGrad, mxnet::op::batchnorm::kBeta);
- PRT(os, *prop.executor_, BlobVectorType::kAux, op::batchnorm::kMovingMean);
- PRT(os, *prop.executor_, BlobVectorType::kAux, op::batchnorm::kMovingVar);
+ PRT(os, *prop.executor_, BlobVectorType::kAux, mxnet::op::batchnorm::kMovingMean);
+ PRT(os, *prop.executor_, BlobVectorType::kAux, mxnet::op::batchnorm::kMovingVar);
- PRT(os, *prop.executor_, BlobVectorType::kOutGrad, op::batchnorm::kOut);
+ PRT(os, *prop.executor_, BlobVectorType::kOutGrad, mxnet::op::batchnorm::kOut);
}
return *os;
}
@@ -625,7 +625,7 @@ static test::op::OpInfoPair<OperatorProp1, OperatorProp2, OperatorExecutor> test
BatchNormValidator<DType, AccReal>::compare(
*info_1.executor_, *info_2.executor_,
test::op::LegacyOperatorExecutor<DType, AccReal>::kInput,
- op::batchnorm::kData, false);
+ mxnet::op::batchnorm::kData, false);
if (!thisCount) {
// return backward
@@ -671,13 +671,14 @@ testForwardAndBackward(const bool isGPU,
}
template<typename OperatorExecutor>
-static test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp, OperatorExecutor>
+static test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp, OperatorExecutor>
testBNForwardAndBackward2D(const bool isGPU,
const TShape &inputShape,
const test::op::kwargs_t kwargs,
const bool dumpC = false) {
CHECK_EQ(inputShape.ndim(), 4); // V1 can only handle 2D
- return testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp, OperatorExecutor>(
+ return testForwardAndBackward<mxnet::op::BatchNormV1Prop,
+ mxnet::op::BatchNormProp, OperatorExecutor>(
isGPU,
isGPU,
inputShape,
@@ -708,14 +709,14 @@ TEST(BATCH_NORM, Test1DForward) {
MSHADOW_REAL_TYPE_SWITCH_EX(
type, DType, AccReal,
{
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
false, {BATCH_SIZE, CHANNELS, DW}, blank_kwargs);
});
}
}
TEST(BATCH_NORM, Test2DForwardV1) {
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<float, float>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<float, float>>(
false,
{BATCH_SIZE, CHANNELS, DH, DW},
blank_kwargs);
@@ -726,7 +727,7 @@ TEST(BATCH_NORM, Test2DForward) {
MSHADOW_REAL_TYPE_SWITCH_EX(
type, DType, AccReal,
{
- auto opInfoFloatH = TestBatchNormOperatorForward<op::BatchNormProp,
+ auto opInfoFloatH = TestBatchNormOperatorForward<mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, {BATCH_SIZE, CHANNELS, DH, DW}, blank_kwargs);
});
@@ -738,7 +739,7 @@ TEST(BATCH_NORM, Test3DForward) {
MSHADOW_REAL_TYPE_SWITCH_EX(
type, DType, AccReal,
{
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
false, {BATCH_SIZE, CHANNELS, DEPTH, DH, DW}, blank_kwargs);
});
}
@@ -829,7 +830,7 @@ TEST(BATCH_NORM, TestStochasticTiming_2D) {
MSHADOW_REAL_TYPE_SWITCH_EX(
mshadow::kFloat32, DType, AccReal,
{
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"RANDOM: BatchNormProp<cpu>", false, true,
blank_kwargs_nocudnn, GPU_TEST_DIMENSIONS); });
#if MXNET_USE_CUDA
@@ -837,7 +838,7 @@ TEST(BATCH_NORM, TestStochasticTiming_2D) {
MSHADOW_REAL_TYPE_SWITCH_EX(
mshadow::kFloat32, DType, AccReal,
{
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"RANDOM: BatchNormProp<gpu>", true, true,
blank_kwargs_nocudnn, GPU_TEST_DIMENSIONS); });
}
@@ -857,38 +858,38 @@ TEST(BATCH_NORM, TestTiming_2D) {
}
MSHADOW_REAL_TYPE_SWITCH_EX(
mshadow::kFloat32, DType, AccReal, {
- timingTest<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
"BatchNormV1Prop<cpu> 2D",
false, false,
blank_kwargs,
2, THISCOUNT);
#if defined(MXNET_USE_MKL2017) && (MXNET_USE_MKL2017 == 1)
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"MKL BatchNormProp<cpu> 2D",
false, false,
blank_kwargs_nocudnn,
2, THISCOUNT);
#endif
test::ScopeSet<volatile bool> disableMKL(&mxnet::op::batchnorm::disable_mkl, true);
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"BatchNormProp<cpu> 2D",
false, false,
blank_kwargs_nocudnn,
2, THISCOUNT);
#if MXNET_USE_CUDA
if (test::unitTestsWithCuda) {
- timingTest<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
"BatchNormV1Prop<gpu> 2D",
true, false,
blank_kwargs,
2, THISCOUNT);
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"BatchNormProp<gpu> 2D",
true, false,
blank_kwargs_nocudnn,
2, THISCOUNT);
#if MXNET_USE_CUDNN == 1 && CUDNN_MAJOR >= 5
- timingTest<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ timingTest<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
"CUDNN BatchNormProp<gpu> 2D",
true, false,
blank_kwargs,
@@ -906,8 +907,8 @@ MSHADOW_REAL_TYPE_SWITCH_EX(
template<typename DType, typename AccReal>
struct BothInfo {
- test::op::OpInfo<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>> info_v1_;
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_;
+ test::op::OpInfo<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>> info_v1_;
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_;
};
TEST(BATCH_NORM, TestBackward2D_Simple) {
@@ -916,9 +917,10 @@ TEST(BATCH_NORM, TestBackward2D_Simple) {
{
const TShape inputShape({1, 1, 2, 1});
test::op::OpInfoPair<
- op::BatchNormV1Prop, op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> bi =
+ mxnet::op::BatchNormV1Prop,
+ mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> bi =
testForwardAndBackward<
- op::BatchNormV1Prop, op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
false, inputShape, blank_kwargs); // Keep it simple
});
}
@@ -946,15 +948,17 @@ TEST(BATCH_NORM, TestIterAll) {
MSHADOW_REAL_TYPE_SWITCH_EX(
type, DType, AccReal,
{
- test::op::OpInfoPair<op::BatchNormProp, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>
- bi = testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ bi = testForwardAndBackward<mxnet::op::BatchNormProp,
+ mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
g1 != 0, g2 != 0, shape, kwargs, false); // Keep it simple
if (shape.ndim() == 4 && type == mshadow::kFloat32 && !x3) {
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>
- bi = testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ bi = testForwardAndBackward<mxnet::op::BatchNormV1Prop,
+ mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
g1 != 0, g2 != 0, shape, kwargs, false); // Keep it simple
}
@@ -985,13 +989,13 @@ static void test_V1_V2_2D(const test::op::kwargs_t &kwargs, const size_t count)
TShape shapes[2] = {2, 3};
const TShape inputShape({2, 3});
- test::op::OpInfo<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>> info_1 =
+ test::op::OpInfo<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>> info_1 =
test::op::createOpAndInfoF<
- op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
+ mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
kwargs, gpu_V1, inputShape);
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_2 =
- test::op::createOpAndInfoF<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_2 =
+ test::op::createOpAndInfoF<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
kwargs, gpu_V2, inputShape);
info_1.executor_->initForward(*info_1.prop_, &info_1.in_type_);
@@ -999,7 +1003,7 @@ static void test_V1_V2_2D(const test::op::kwargs_t &kwargs, const size_t count)
info_1.executor_->initBackward(*info_1.prop_, &info_1.in_type_);
info_2.executor_->initBackward(*info_1.prop_, &info_1.in_type_);
- TBlob &blob1 = info_1.executor_->inputs()[op::batchnorm::kData];
+ TBlob &blob1 = info_1.executor_->inputs()[mxnet::op::batchnorm::kData];
test::data_ref<DType>(&blob1, {0, 0}) = -0.05f;
test::data_ref<DType>(&blob1, {0, 1}) = -0.19f;
test::data_ref<DType>(&blob1, {0, 2}) = 0.02f;
@@ -1007,7 +1011,7 @@ static void test_V1_V2_2D(const test::op::kwargs_t &kwargs, const size_t count)
test::data_ref<DType>(&blob1, {1, 1}) = 0.06f;
test::data_ref<DType>(&blob1, {1, 2}) = -0.01f;
- TBlob &blob2 = info_2.executor_->inputs()[op::batchnorm::kData];
+ TBlob &blob2 = info_2.executor_->inputs()[mxnet::op::batchnorm::kData];
test::data_ref<DType>(&blob2, {0, 0}) = -0.05f;
test::data_ref<DType>(&blob2, {0, 1}) = -0.19f;
test::data_ref<DType>(&blob2, {0, 2}) = 0.02f;
@@ -1015,11 +1019,11 @@ static void test_V1_V2_2D(const test::op::kwargs_t &kwargs, const size_t count)
test::data_ref<DType>(&blob2, {1, 1}) = 0.06f;
test::data_ref<DType>(&blob2, {1, 2}) = -0.01f;
- test::data_ref<DType>(&info_1.executor_->inputs()[op::batchnorm::kGamma], {1}) = 3;
- test::data_ref<DType>(&info_2.executor_->inputs()[op::batchnorm::kGamma], {1}) = 3;
+ test::data_ref<DType>(&info_1.executor_->inputs()[mxnet::op::batchnorm::kGamma], {1}) = 3;
+ test::data_ref<DType>(&info_2.executor_->inputs()[mxnet::op::batchnorm::kGamma], {1}) = 3;
- test::data_ref<DType>(&info_1.executor_->inputs()[op::batchnorm::kBeta], {0}) = 3;
- test::data_ref<DType>(&info_2.executor_->inputs()[op::batchnorm::kBeta], {0}) = 3;
+ test::data_ref<DType>(&info_1.executor_->inputs()[mxnet::op::batchnorm::kBeta], {0}) = 3;
+ test::data_ref<DType>(&info_2.executor_->inputs()[mxnet::op::batchnorm::kBeta], {0}) = 3;
for (size_t x = 0; x < count; ++x) {
info_1.executor_->forward();
@@ -1052,9 +1056,9 @@ TEST(BATCH_NORM, TestBackward2D_SimpleNFG) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, inputShape, nonfixgamma_kwargs);
});
@@ -1066,9 +1070,9 @@ TEST(BATCH_NORM, Test2DBackward_Complex) {
{
test::ScopeSet<bool> noDebugOutput(&test::debug_output, false);
const TShape inputShape({9, 14, 16, 91});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, inputShape, blank_kwargs);
});
@@ -1078,9 +1082,9 @@ struct Test2DBackward2DPlusLoadAndCompareLogicUtil {
template <typename DType, typename AccReal>
static void test() {
const TShape inputShape({1, 1, 2, 1});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, inputShape, blank_kwargs, false, 1, 5);
@@ -1116,8 +1120,8 @@ struct Test2DBackward2DPlusLoadAndCompareLogicUtil {
// Expected data state when running forward+backward starting with default values
// Note: This data structure generated by dumpC()
// Test loaded data agsinst calculated data
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_checkLoad =
- test::op::createOpAndInfoF<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_checkLoad =
+ test::op::createOpAndInfoF<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
blank_kwargs, false, inputShape);
info_checkLoad.executor_->initForward(*info_checkLoad.prop_, &info_checkLoad.in_type_);
info_checkLoad.executor_->initBackward(*info_checkLoad.prop_, &info_checkLoad.in_type_);
@@ -1159,8 +1163,8 @@ TEST(BATCH_NORM, TestBackward1D_Simple) {
mshadow::kFloat32, DTypeX, AccReal,
{
const TShape inputShape({1, 1, 2});
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DTypeX, AccReal>> info =
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DTypeX, AccReal>>(
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DTypeX, AccReal>> info =
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DTypeX, AccReal>>(
false, inputShape, blank_kwargs);
info.executor_->initBackward(*info.prop_, &info.in_type_);
runOperatorBackward(&info);
@@ -1207,8 +1211,8 @@ TEST(BATCH_NORM, TestBackward3D) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({2, 3, 2, 3, 5});
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info =
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info =
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
false, inputShape, blank_kwargs);
info.executor_->initBackward(*info.prop_, &info.in_type_);
runOperatorBackward(&info);
@@ -1225,9 +1229,9 @@ TEST(BATCH_NORM, Test2DBackwardMixed_cpu_cpu_nfg) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, false, inputShape, nonfixgamma_kwargs, false);
dumpF(&std::cout, bi);
@@ -1241,9 +1245,9 @@ TEST(BATCH_NORM, Test2DBackwardMixed_cpu_cpu_ugs) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, false, inputShape, useglobalstats_kwargs, false);
dumpF(&std::cout, bi);
@@ -1504,16 +1508,16 @@ static void runChannelAxisTest(
// Create operator 1 with ChannelAxis2 (normally the experimental one)
kwargs.push_back({"axis", std::to_string(channelAxis1)});
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_c1 =
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_c1 =
test::op::createOpAndInfoF<
- op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
kwargs, isGPU1, shape_c1);
// Create operator 2 with ChannelAxis2 (normally the control one)
kwargs.pop_back();
kwargs.push_back({"axis", std::to_string(channelAxis2)});
- test::op::OpInfo<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_c2 =
- test::op::createOpAndInfoF<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ test::op::OpInfo<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>> info_c2 =
+ test::op::createOpAndInfoF<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
kwargs, isGPU2, shape_c2);
kwargs.pop_back();
@@ -1643,11 +1647,11 @@ TEST(BATCH_NORM, Test2DForwardV12D_gpu) {
MSHADOW_REAL_TYPE_SWITCH_EX(
mshadow::kFloat32, DType, AccReal,
{
- TestBatchNormOperatorForward<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
true,
{BATCH_SIZE, CHANNELS, DH, DW},
blank_kwargs);
- TestBatchNormOperatorForward<op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormV1Prop, BNOperatorExecutor<DType, AccReal>>(
true,
{BATCH_SIZE, CHANNELS, DH, DW},
blank_kwargs);
@@ -1659,11 +1663,11 @@ TEST(BATCH_NORM, Test2DForward2D_gpu) {
MSHADOW_REAL_TYPE_SWITCH_EX(
type, DType, AccReal,
{
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
true,
{BATCH_SIZE, CHANNELS, DH, DW},
blank_kwargs);
- TestBatchNormOperatorForward<op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
+ TestBatchNormOperatorForward<mxnet::op::BatchNormProp, BNOperatorExecutor<DType, AccReal>>(
true,
{BATCH_SIZE, CHANNELS, DH, DW},
blank_kwargs_nocudnn);
@@ -1677,7 +1681,7 @@ TEST(BATCH_NORM, Test2DBackwardMixedV1_gpu_cpu) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormV1Prop,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormV1Prop,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs, false);
});
@@ -1688,7 +1692,7 @@ TEST(BATCH_NORM, Test2DBackwardMixedV1Complex_gpu_cpu) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormV1Prop,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormV1Prop,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs, false);
});
@@ -1700,10 +1704,10 @@ TEST(BATCH_NORM, Test2DBackwardMixed_gpu_cpu) {
type, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs_nocudnn, false);
});
@@ -1716,10 +1720,10 @@ TEST(BATCH_NORM, Test2DBackwardMixedComplex_gpu_cpu) {
type, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, blank_kwargs_nocudnn, false);
});
@@ -1734,7 +1738,7 @@ TEST(BATCH_NORM, Test2DBackwardMixedV1V2Complex_cpu_cpu_nfg) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, false, inputShape, nonfixgamma_kwargs, false);
});
@@ -1746,10 +1750,10 @@ TEST(BATCH_NORM, Test2DBackwardMixed_gpu_cpu_nfg) {
type, DType, AccReal,
{
const TShape inputShape({1, 1, 2, 1});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, nonfixgamma_kwargs, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, nonfixgamma_kwargs_nocudnn, false);
});
@@ -1762,10 +1766,10 @@ TEST(BATCH_NORM, Test2DBackwardMixedComplex_gpu_cpu_nfg) {
type, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, nonfixgamma_kwargs, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, nonfixgamma_kwargs_nocudnn, false);
});
@@ -1780,9 +1784,9 @@ TEST(BATCH_NORM, Test2DBackwardMixedV1V2Complex_cpu_cpu_ugs) {
mshadow::kFloat32, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- test::op::OpInfoPair<op::BatchNormV1Prop, op::BatchNormProp,
+ test::op::OpInfoPair<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>> bi =
- testForwardAndBackward<op::BatchNormV1Prop, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormV1Prop, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, false, inputShape, useglobalstats_kwargs, false);
dumpF(&std::cout, bi);
@@ -1796,10 +1800,10 @@ TEST(BATCH_NORM, Test2DBackwardMixed_gpu_cpu_ugs) {
type, DType, AccReal,
{
const TShape inputShape({2, 3, 2, 2});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, useglobalstats_kwargs_nocudnn, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, useglobalstats_kwargs, false);
});
@@ -1812,10 +1816,10 @@ TEST(BATCH_NORM, Test2DBackwardMixedComplex_gpu_cpu_ugs) {
type, DType, AccReal,
{
const TShape inputShape({BATCH_SIZE, CHANNELS, DH, DW});
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, useglobalstats_kwargs, false);
- testForwardAndBackward<op::BatchNormProp, op::BatchNormProp,
+ testForwardAndBackward<mxnet::op::BatchNormProp, mxnet::op::BatchNormProp,
BNOperatorExecutor<DType, AccReal>>(
false, true, inputShape, useglobalstats_kwargs_nocudnn, false);
});
diff --git a/tests/cpp/operator/tune/operator_tune_test.cc b/tests/cpp/operator/tune/operator_tune_test.cc
index 60096d33df..f404e4faa9 100644
--- a/tests/cpp/operator/tune/operator_tune_test.cc
+++ b/tests/cpp/operator/tune/operator_tune_test.cc
@@ -31,7 +31,8 @@ using namespace mxnet;
* \brief ActivationOp timing test for CPU
*/
TEST(OMP_TUNING, ShowAllTunedOps) {
- const std::unordered_set<std::string>& op_names = op::OperatorTune<float>::TunedOperatorNames();
+ const std::unordered_set<std::string>& op_names =
+ mxnet::op::OperatorTune<float>::TunedOperatorNames();
for (auto iter = op_names.begin(), e_iter = op_names.end(); iter != e_iter; ++iter) {
std::cout << *iter << std::endl;
}
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
With regards,
Apache Git Services