You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by li...@apache.org on 2020/01/27 18:28:06 UTC

[incubator-mxnet] 02/02: upgrade enum according to updated tvm

This is an automated email from the ASF dual-hosted git repository.

liuyizhi pushed a commit to branch tvm_sync
in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git

commit 2ef7de0ec0072828e788976d1ec44e9438b96383
Author: Yizhi Liu <li...@apache.org>
AuthorDate: Fri Jan 24 22:17:50 2020 -0800

    upgrade enum according to updated tvm
---
 src/nnvm/plan_memory.cc                              | 2 --
 src/nnvm/tvm_bridge.cc                               | 4 ++--
 src/operator/numpy/np_elemwise_broadcast_logic_op.cc | 6 +++---
 src/operator/tensor/elemwise_unary_op_pow.cc         | 4 ++--
 src/operator/tvmop/op_module.cc                      | 2 +-
 5 files changed, 8 insertions(+), 10 deletions(-)

diff --git a/src/nnvm/plan_memory.cc b/src/nnvm/plan_memory.cc
index c89eefc..e061dab 100644
--- a/src/nnvm/plan_memory.cc
+++ b/src/nnvm/plan_memory.cc
@@ -26,7 +26,6 @@
 #include <nnvm/pass.h>
 #include <nnvm/graph_attr_types.h>
 #include <nnvm/op_attr_types.h>
-#include <nnvm/top/tensor.h>
 #include <mxnet/base.h>
 #include <memory>
 #include "graph_algorithm.h"
@@ -36,7 +35,6 @@ namespace nnvm {
 namespace pass {
 
 namespace {
-  using namespace nnvm::top;
 // Return bytes of data flag.
 static int MXGetDTypeSize(int type_flag) {
   switch (type_flag) {
diff --git a/src/nnvm/tvm_bridge.cc b/src/nnvm/tvm_bridge.cc
index 0692998..17e05e3 100644
--- a/src/nnvm/tvm_bridge.cc
+++ b/src/nnvm/tvm_bridge.cc
@@ -73,7 +73,7 @@ class TVMFunctor {
         const NDArray& nd =
             static_cast<NDArray*>(args.values[i].v_handle)[0];
         // We cannot set the value until
-        type_codes_[i] = kArrayHandle;
+        type_codes_[i] = kTVMDLTensorHandle;
         array_data_.push_back(nd);
         array_loc_.push_back(i);
         // check if there is read or mutate
@@ -86,7 +86,7 @@ class TVMFunctor {
           mutate_vars->push_back(nd.var());
         }
       } else {
-        CHECK_LT(args.type_codes[i], kTVMType)
+        CHECK_LT(args.type_codes[i], kTVMDataType)
             << "Only allow POD type in mxnet async call";
       }
     }
diff --git a/src/operator/numpy/np_elemwise_broadcast_logic_op.cc b/src/operator/numpy/np_elemwise_broadcast_logic_op.cc
index 7e8951a..8395caf 100644
--- a/src/operator/numpy/np_elemwise_broadcast_logic_op.cc
+++ b/src/operator/numpy/np_elemwise_broadcast_logic_op.cc
@@ -95,7 +95,7 @@ struct TVMBinaryBroadcastCompute {
     values.resize(num_args);
     for (size_t i = 0; i < num_args; ++i) {
       tblobs[i] = PrependAxes(tblobs[i], ondim);
-      type_codes[i] = kArrayHandle;
+      type_codes[i] = kTVMDLTensorHandle;
       values[i].v_handle = const_cast<DLTensor*>(&(tblobs[i].dltensor()));
     }
     tvm::runtime::TVMArgs tvm_args(&values[0], &type_codes[0], tblobs.size());
@@ -200,7 +200,7 @@ struct TVMBinaryBroadcastScalarCompute {
     values.resize(num_args);
 
     // input tensor setup
-    type_codes[0] = kArrayHandle;
+    type_codes[0] = kTVMDLTensorHandle;
     values[0].v_handle = const_cast<DLTensor*>(&(tblobs[0].dltensor()));
 
     // scalar param
@@ -208,7 +208,7 @@ struct TVMBinaryBroadcastScalarCompute {
     values[1].v_float64 = nnvm::get<double>(attrs.parsed);
 
     // output tensor
-    type_codes[2] = kArrayHandle;
+    type_codes[2] = kTVMDLTensorHandle;
     values[2].v_handle = const_cast<DLTensor*>(&(tblobs[1].dltensor()));
 
     tvm::runtime::TVMArgs tvm_args(&values[0], &type_codes[0], 3);
diff --git a/src/operator/tensor/elemwise_unary_op_pow.cc b/src/operator/tensor/elemwise_unary_op_pow.cc
index b4d3a4a..914cb820 100644
--- a/src/operator/tensor/elemwise_unary_op_pow.cc
+++ b/src/operator/tensor/elemwise_unary_op_pow.cc
@@ -224,7 +224,7 @@ The storage type of ``rsqrt`` output is always dense
 MXNET_OPERATOR_REGISTER_BINARY_WITH_SPARSE_CPU_DR(
   _backward_rsqrt, unary_bwd<mshadow_op::reciprocal_square_root_grad>)
 .set_attr<nnvm::FGradient>("FGradient",
-  [](const nnvm::NodePtr& n, const std::vector<nnvm::NodeEntry>& ograds) {
+  [](const nnvm::ObjectPtr& n, const std::vector<nnvm::NodeEntry>& ograds) {
       // NodeEntry{n} : y_grad * f'(x)
       // n->inputs[0] : y_grad
       // n->inputs[1] : x
@@ -329,7 +329,7 @@ MXNET_OPERATOR_REGISTER_BINARY(_backward_rcbrt)
                     ElemwiseBinaryOp::Compute<cpu,
                       unary_bwd<mshadow_op::reciprocal_cube_root_grad>>)
 .set_attr<nnvm::FGradient>("FGradient",
-  [](const nnvm::NodePtr& n, const std::vector<nnvm::NodeEntry>& ograds) {
+  [](const nnvm::ObjectPtr& n, const std::vector<nnvm::NodeEntry>& ograds) {
       // NodeEntry{n} : y_grad * f'(x)
       // n->inputs[0] : y_grad
       // n->inputs[1] : x
diff --git a/src/operator/tvmop/op_module.cc b/src/operator/tvmop/op_module.cc
index b45df5d..cdd7321 100644
--- a/src/operator/tvmop/op_module.cc
+++ b/src/operator/tvmop/op_module.cc
@@ -94,7 +94,7 @@ void TVMOpModule::Call(const std::string &func_name,
   type_codes.resize(args.size());
   values.resize(args.size());
   for (size_t i = 0; i < args.size(); ++i) {
-    type_codes[i] = kArrayHandle;
+    type_codes[i] = kTVMDLTensorHandle;
     values[i].v_handle = const_cast<DLTensor *>(&(args[i].dltensor()));
   }