tf.function not used for model inference #10728
Labels
area:rasa-oss 🎡
Anything related to the open source Rasa framework
area:rasa-oss/performance
Issues focused around the performance of `rasa` (e.g.number of handled requests per second)
type:bug 🐛
Inconsistencies or issues which will cause an issue or problem for users or implementors.
Rasa Open Source version
3.0.x
Rasa SDK version
No response
Rasa X version
No response
Python version
3.8
What operating system are you using?
OSX
What happened?
When running model inference in
RasaModel
, the predict step being used here is the (eager) Python implementation, not thetf.function
created just before that. Therefore, the Tensorflow graph is not used for performance optimization.Testing a quick fix on the Sara dataset with config from here (limiting training epochs to 20 steps) and timing
gave the following inference improvements (tested locally on Macbook Pro 2,4 GHz 8-Core Intel Core i9):
Log output:
test-main.log
test-fix.log
Command / Request
Definition of Done
The text was updated successfully, but these errors were encountered: