You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@tvm.apache.org by GitBox <gi...@apache.org> on 2022/09/13 06:52:33 UTC
[GitHub] [tvm] billishyahao opened a new issue, #12763: [Bug][DNNL][BYOC] Transform Pass From float32 to bfloat16 failed in some patterns
billishyahao opened a new issue, #12763:
URL: https://github.com/apache/tvm/issues/12763
When we invoke
```
relay_mod = relay.transform.ToMixedPrecision('bfloat16')(relay_mod)
```
pass, we are supposed to get a well-transformed model which contains different kinds of operators with auto mix precision.
However, we have seen some error messages.
```
The Relay type checker is unable to show the following types match.
In particular `Tensor[(64), float32]` does not match `Tensor[(64), bfloat16]`
The Relay type checker is unable to show the following types match.
In particular `Tensor[(64), float32]` does not match `Tensor[(64), bfloat16]`
```
After we print out the relay graph, we can see this error occurs because nn.layer_norm belongs to DEFAULT_NEVER_LIST but there is not cast operators which promoted its input from bfloat16 to float32.
There is a minimal script which can reproduce this issue:
```
def test_layernorm(run_module, dtype="float32"):
x = relay.var("x", shape=(4, 3, 224, 224), dtype="float32")
k_shape = (64, 3, 4, 4)
kernel = relay.const(np.random.randint(0, 1, k_shape).astype(dtype))
out = relay.nn.conv2d(
x,
kernel,
channels=64,
kernel_size=k_shape[2:4],
strides=[4, 4],
padding=[0, 0, 0, 0]
)
# bias = relay.var("bias", shape=(64), dtype=dtype)
bias = relay.const(np.random.randint(0, 1, (64,)).astype(dtype))
out = relay.nn.bias_add(out, bias)
out = relay.reshape(out, newshape=[4, 64, -1])
out = relay.transpose(out, axes=[0, 2, 1])
beta = relay.const(np.zeros((64)).astype("float32"))
gamma = relay.const(np.ones((64)).astype("float32"))
out = relay.nn.layer_norm(out, gamma=gamma, beta=beta)
dic = {"x": (4, 3, 224, 224), "kernel": k_shape, "bias": (4, 64, 56, 56)}
param_lst = ["kernel", "bias"]
out = tvm.IRModule.from_expr(out)
config = out, dic, param_lst
run_and_verify_func(config, run_module=run_module, dtype=dtype)
```
I posted the code into my personal repo:
https://github.com/apache/tvm/commit/fa1fb3ba1854c8c2c6fa813f95bb2454ff2f249c#diff-8b8dff568d8ee66e899bdc2b00d56db2a75e050c42b771fe38df17c4ca1ccdfb
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@tvm.apache.org.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
[GitHub] [tvm] yangulei commented on issue #12763: [Bug][DNNL][BYOC] Transform Pass From float32 to bfloat16 failed in some patterns
Posted by GitBox <gi...@apache.org>.
yangulei commented on issue #12763:
URL: https://github.com/apache/tvm/issues/12763#issuecomment-1247463034
This bug will be fixed by https://github.com/apache/tvm/pull/12787, please help to review the PR, thanks.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@tvm.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
[GitHub] [tvm] masahi closed issue #12763: [Bug][DNNL][BYOC] Transform Pass From float32 to bfloat16 failed in some patterns
Posted by GitBox <gi...@apache.org>.
masahi closed issue #12763: [Bug][DNNL][BYOC] Transform Pass From float32 to bfloat16 failed in some patterns
URL: https://github.com/apache/tvm/issues/12763
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@tvm.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org