diff --git a/spackbot/handlers/mirrors.py b/spackbot/handlers/mirrors.py index 40b77fb..4b08e13 100644 --- a/spackbot/handlers/mirrors.py +++ b/spackbot/handlers/mirrors.py @@ -5,15 +5,9 @@ import spackbot.helpers as helpers from spackbot.helpers import ( - pr_expected_base, pr_mirror_base_url, - pr_shared_mirror, - publish_mirror_base_url, ) from spackbot.workers import ( - copy_pr_mirror, - prune_mirror_duplicates, - update_mirror_index, delete_pr_mirror, get_queue, TASK_QUEUE_LONG, @@ -33,69 +27,19 @@ async def close_pr_mirror(event, gh): return # Get PR event info - base_branch = payload["pull_request"]["base"]["ref"] - is_merged = payload["pull_request"]["merged"] pr_number = payload["number"] pr_branch = payload["pull_request"]["head"]["ref"] pr_mirror_url = f"{pr_mirror_base_url}/pr{pr_number}_{pr_branch}" - shared_pr_mirror_url = f"{pr_mirror_base_url}/{pr_shared_mirror}" # Get task queue info ltask_q = get_queue(TASK_QUEUE_LONG) - copy_job = None job_metadata = { "type": None, "pr_number": pr_number, "pr_branch": pr_branch, } - if is_merged and base_branch == pr_expected_base: - logger.info( - f"PR {pr_number}/{pr_branch} merged to develop, graduating binaries" - ) - - # Copy all of the stack binaries from the PR to the shared PR - # mirror. - job_metadata.update({"type": "copy"}) - copy_job = ltask_q.enqueue( - copy_pr_mirror, - pr_mirror_url, - shared_pr_mirror_url, - meta=job_metadata, - job_timeout=WORKER_JOB_TIMEOUT, - ) - logger.info(f"Copy job queued: {copy_job.id}") - - # Prune duplicates that have been published after copy - # since copy may have introduced duplicates for some reason - job_metadata.update({"type": "prune"}) - shared_stack_pr_mirror_url = f"{shared_pr_mirror_url}/{{stack}}" - publish_stack_mirror_url = ( - f"{publish_mirror_base_url}/{{stack}}/{pr_expected_base}" - ) - prune_job = ltask_q.enqueue( - prune_mirror_duplicates, - shared_stack_pr_mirror_url, - publish_stack_mirror_url, - job_timeout=WORKER_JOB_TIMEOUT, - depends_on=copy_job, - meta=job_metadata, - ) - logger.info(f"Pruning job queued: {prune_job.id}") - - # Queue a reindex for the stack mirror to attempt to run after - # prune. - job_metadata.update({"type": "reindex"}) - update_job = ltask_q.enqueue( - update_mirror_index, - shared_stack_pr_mirror_url, - job_timeout=WORKER_JOB_TIMEOUT, - depends_on=prune_job, - meta=job_metadata, - ) - logger.info(f"Reindex job queued: {update_job.id}") - # Delete the mirror job_metadata.update({"type": "delete"}) del_job = ltask_q.enqueue( @@ -103,6 +47,5 @@ async def close_pr_mirror(event, gh): pr_mirror_url, meta=job_metadata, job_timeout=WORKER_JOB_TIMEOUT, - depends_on=copy_job, ) logger.info(f"Delete job queued: {del_job.id}")