Skip to content
This repository has been archived by the owner on Oct 27, 2023. It is now read-only.

Commit

Permalink
Adjust api.yaml
Browse files Browse the repository at this point in the history
  • Loading branch information
atuchin-m committed Oct 25, 2023
1 parent 0d9ea47 commit 09eb3be
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 8 deletions.
8 changes: 4 additions & 4 deletions dashboard/api.yaml
Original file line number Diff line number Diff line change
@@ -1,20 +1,20 @@
service: api

runtime: python38
entrypoint: gunicorn -b:$PORT dashboard.dispatcher:APP --worker-class gthread --threads 10 --timeout 60
entrypoint: gunicorn -b:$PORT dashboard.dispatcher:APP -w 4 --timeout 60
app_engine_apis: true
# 2048MB Memory, 2.4GHz CPU limit, support auto scaling.
instance_class: F4_1G
instance_class: F2

automatic_scaling:
# We're setting the max concurrent request to 20, to allow AppEngine to scale
# the number of instances to handle API requests better. We're also going to
# keep around 10 instances at the ready to handle incoming requests better
# from a "cold start".
max_concurrent_requests: 20
max_concurrent_requests: 8
max_pending_latency: automatic
min_instances: 0
max_instances: 3
max_instances: 6
target_cpu_utilization: 0.8

env_variables:
Expand Down
8 changes: 4 additions & 4 deletions dashboard/app.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,14 +3,14 @@

runtime: python38

entrypoint: gunicorn -b:$PORT dashboard.dispatcher:APP --worker-class gthread --threads 10 --timeout 60
entrypoint: gunicorn -b:$PORT dashboard.dispatcher:APP -w 4 --timeout 60
app_engine_apis: true
# 2048MB Memory, 2.4GHz CPU limit, support auto scaling.
instance_class: F4_1G
instance_class: F2

automatic_scaling:
max_concurrent_requests: 80
max_instances: 2
max_concurrent_requests: 40
max_instances: 6
max_pending_latency: automatic
min_instances: 0
target_cpu_utilization: 0.8
Expand Down

0 comments on commit 09eb3be

Please sign in to comment.