You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@tvm.apache.org by GitBox <gi...@apache.org> on 2022/05/06 20:03:40 UTC

[GitHub] [tvm] Wheest commented on pull request #11231: Update debugger.rst

Wheest commented on PR #11231:
URL: https://github.com/apache/tvm/pull/11231#issuecomment-1119970240

   That's correct @tkonolige, [this Python script](https://gist.github.com/Wheest/9ad2d6a47bbd2cfaa4be530c68ba2f6c) shows the issue.
   
   The script has three modes: `["tutorial", "alt", "normal"]`, invoked with `python tvm_profiler_simple.py --mode tutorial`.
   
   - `normal` inference works of course
   - `tutorial` is the approach in the current documentation, with fails with output [1]
   - `alt` is the approach in this PR, which works, with example output [2]
   
   As the [discussion in the forum says](https://discuss.tvm.apache.org/t/runnig-a-model-with-tvm-debugger/9869/9?u=wheest), there does not appear to be a canonical way of creating the debugger, but the approach in the docs should at least work, until there has been a refactor to produce an unambiguous canonical approach.  
   
   
   
   [1] `tutorial` sample output:
   ```
   Traceback (most recent call last):
     File "tvm_profiler_simple.py", line 98, in <module>
       main(args)
     File "tvm_profiler_simple.py", line 64, in main
       m = debug_executor.create(lib.graph_json, lib, dev, dump_root="/tmp/tvmdbg")
     File "/app/source/tvm/python/tvm/contrib/debugger/debug_executor.py", line 70, in create
       func_obj = fcreate(graph_json_str, libmod, *device_type_id)
     File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 223, in __call__
       values, tcodes, num_args = _make_tvm_args(args, temp_args)
     File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 188, in _make_tvm_args
       raise TypeError("Don't know how to handle type %s" % type(arg))
   TypeError: Don't know how to handle type <class 'tvm.relay.backend.executor_factory.GraphExecutorFactoryModule'>
   
   ```
   
   [2] `alt` sample output
   ```
   [19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:103: Iteration: 0
   [19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:108: Op #0 tvmgen_default_fused_layout_transform: 29.3538 us/iter
   [19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:108: Op #1 tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu: 3009 us/iter
   ...
   Node Name                                                       Ops                                                             Time(us)   Time(%)  Shape                 Inputs  Outputs
   ---------                                                       ---                                                             --------   -------  -----                 ------  -------
   tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu        tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu        3009.0     7.115    (1, 2, 112, 112, 32)  3       1
   tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_add_nn_relu_4  tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_add_nn_relu_4  2835.26    6.704    (1, 16, 14, 14, 16)   4       1
   ...
   ```
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@tvm.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org