From 3056b4f6176d25821ddc49a57ac8bb9b2afaaf0f Mon Sep 17 00:00:00 2001 From: Nikolai Kondrashov Date: Wed, 20 Nov 2024 13:49:04 +0200 Subject: [PATCH] cloud: Bump archival RAM to (the max of) 8GB Bump the RAM allocated to the data archival cloud function from 4GB to 8GB, so we avoid aborting on out-of-memory when transferring 12 hours of data at a time. It could be that BigQuery is accumulating memory use over time and running out, despite us calling gc.collect() after every transfer, but I don't have the time to pursue this further. --- kcidb/cloud/functions.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/kcidb/cloud/functions.sh b/kcidb/cloud/functions.sh index f47053ac..7b103f41 100644 --- a/kcidb/cloud/functions.sh +++ b/kcidb/cloud/functions.sh @@ -181,7 +181,7 @@ function functions_deploy() { archive true \ --env-vars-file "$env_yaml_file" \ --trigger-topic "${archive_trigger_topic}" \ - --memory 4096MB \ + --memory 8192MB \ --max-instances=1 \ --timeout 540