Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Optimize the inference performance on Cloud Run #237

Open
Aniket-Parlikar opened this issue Nov 8, 2022 · 2 comments
Open

Optimize the inference performance on Cloud Run #237

Aniket-Parlikar opened this issue Nov 8, 2022 · 2 comments
Assignees

Comments

@Aniket-Parlikar
Copy link
Contributor

We are currently using Google Cloud Run to obtain inferences on the map files generated using the Google Earth Engine API. However, for sending in a large number of concurrent requests(>175) we are experiencing a decline in performance of the Cloud Run services. Some of the indicators of such performance consist of

  • request latencies,.
  • CPU and memory utilization.
  • container startup time.
  • Total inference time of all the concurrent requests.

While deploying a service on the Google Cloud Run, we need to configure the following parameters( Container CPU and Memory, Request time-out and # of concurrent requests to be sent.)

We have developed a python utility(src/inference_profiling/profile_program.py) to send desired number of concurrent requests to the Cloud Run API and obtain total inference time for those requests.

Inference_issue

Hence, can we get a better understanding of the performance by tuning the the Cloud Run configuration parameters and also be able to optimize it?

@Aniket-Parlikar Aniket-Parlikar self-assigned this Nov 8, 2022
@Aniket-Parlikar
Copy link
Contributor Author

@ivanzvonkov can you please let me know if this looks good or anything needs to be added

@ivanzvonkov
Copy link
Collaborator

ivanzvonkov commented Nov 8, 2022

@Aniket-Parlikar this looks good! How does #201 relate and help with this?
Thanks for all the detail!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants