Mixed precision and Tensorflow XLA #322
-
It's indispensable that we could make all models be able to take advantage of |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 4 replies
-
Thanks @dathudeptrai, we really value your input! We are currently fleshing out our core API but will be emphasizing both performance and usability in our development. Would you be interested in talking to us about your use case so we can learn more about potential users? |
Beta Was this translation helpful? Give feedback.
-
@jbischof My use case is just a general use case where I always try to maximize tf performance not only in training but also in deployment (sever and mobile). Below are some important notes:
Given all points above, I believe if this framework can support all cool features of TF2 including but not limited to mixed precision, xla, pruning, sparsity, tensorrt, irree ... then this framework will lead the market, at least those who do the product and are more prone to deployment. Also note that, the industry community is much larger than the research community. cc @bhack @LukeWood @ianstenbit from |
Beta Was this translation helpful? Give feedback.
-
Thanks for opening! I think this is a great callout! We are just progressing to covering modeling workflows in KerasNLP, but I think what you are saying is spot on. We should really strive to have mixed-precision and XLA not just supported, but enabled as the default option wherever possible. They are both huge speedups. For XLA, I believe that most components where we need support either have it, or there is work happening to support it (e.g. beam search utility). For mixed precision, we do need to figure out how to instantiate say a BERT network with mixed precision along with checkpointed weights. Looks like @jbischof just opened a bug, so will kick off a comment there. #323 |
Beta Was this translation helpful? Give feedback.
Thanks for opening! I think this is a great callout!
We are just progressing to covering modeling workflows in KerasNLP, but I think what you are saying is spot on. We should really strive to have mixed-precision and XLA not just supported, but enabled as the default option wherever possible. They are both huge speedups.
For XLA, I believe that most components where we need support either have it, or there is work happening to support it (e.g. beam search utility).
For mixed precision, we do need to figure out how to instantiate say a BERT network with mixed precision along with checkpointed weights. Looks like @jbischof just opened a bug, so will kick off a comment there. #323