You are viewing a plain text version of this content. The canonical link for it is here.
Posted to discuss-archive@tvm.apache.org by Kylin via Apache TVM Discuss <no...@discuss.tvm.ai> on 2021/04/02 08:05:57 UTC

[Apache TVM Discuss] [Questions] Very slow under linux cuda


Hello~ I am facing the similar problem!

I use autotvm tuned a CNN model trained by tensorflow, all the ops in the model were tuned.

After that, I load log file by relay, and test it’s performence, I found the whole TVM infer time is bigger than the tensofrflow far away.

The "mod.get_out(0).asnumpy()" time is about 240ms!!!

I observe the following information when I test the TVM tuned model.

    Extract tasks...
    Compile...
    Cannot find config for target=cuda -keys=cuda,gpu -max_num_threads=1024 -model=unknown - 
    thread_warp_size=32, workload=('dense_small_batch.cuda', ('TENSOR', (2500, 512), 'float32'), 
    ('TENSOR', (6600, 512), 'float32'), None, 'float32'). A fallback configuration is used, which may bring 
    great performance regression.
How to fix this non-exists configuration for workload named "dense_small_batch.cuda" ?

Looking for your reply! Thank you very much!





---
[Visit Topic](https://discuss.tvm.apache.org/t/very-slow-under-linux-cuda/4793/7) to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click here](https://discuss.tvm.apache.org/email/unsubscribe/d46b4b37769b0764f18db43b48b4dca60cb3fc42855a4ad6f2d59fbf8e9956f4).