-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Speed]speed up python executor in fluid #8729
Comments
Related issue
|
profile with parallel_doone card with parallel_do
conclusionparallel_do add about 40% of performance loss with the program cache without parallel_dowith parallel_do |
The existence of Feed and Fetch operators is a wrong design:
|
Background
problem
In our Python executor, every executor.run will clone the program and then add feed and fetch op to the cloned program. The following profile demonstrates that the Program.clone is very time-consuming. I add a simple cache to void the program clone, and the result is very impressive.
solution
avoid the clone.
Experiment
profile script
#8674
condition
timeline
before optimize
after optimize
The text was updated successfully, but these errors were encountered: