1
0
mirror of https://github.com/quay/quay.git synced 2025-12-19 14:02:30 +03:00
Files
quay/workers/namespacegcworker.py
Joseph Schorr 3f8221f74d GC query improvements (#356)
* Change storage GC to process a single row at a time

This should remove the deadlock under the transaction and be much less
heavy on the DB

* Ensure we don't select repositories for GC from those already marked
for deletion or those under to-be-deleted namespaces

* Ensure that GC operations occur under global locks, to prevent
concurrent GC of the same repositories, which should reduce lock
contention on the database
2020-05-12 11:56:03 -04:00

47 lines
1.4 KiB
Python

import logging
from app import namespace_gc_queue, all_queues
from data import model
from workers.queueworker import QueueWorker
from util.log import logfile_path
from util.locking import GlobalLock, LockNotAcquiredException
logger = logging.getLogger(__name__)
POLL_PERIOD_SECONDS = 60
NAMESPACE_GC_TIMEOUT = 60 * 60 # 60 minutes
LOCK_TIMEOUT_PADDING = 60 # seconds
class NamespaceGCWorker(QueueWorker):
"""
Worker which cleans up namespaces enqueued to be GCed.
"""
def process_queue_item(self, job_details):
try:
with GlobalLock(
"LARGE_GARBAGE_COLLECTION", lock_ttl=NAMESPACE_GC_TIMEOUT + LOCK_TIMEOUT_PADDING
):
self._perform_gc(job_details)
except LockNotAcquiredException:
logger.debug("Could not acquire global lock for garbage collection")
def _perform_gc(self, job_details):
logger.debug("Got namespace GC queue item: %s", job_details)
marker_id = job_details["marker_id"]
model.user.delete_namespace_via_marker(marker_id, all_queues)
if __name__ == "__main__":
logging.config.fileConfig(logfile_path(debug=False), disable_existing_loggers=False)
logger.debug("Starting namespace GC worker")
worker = NamespaceGCWorker(
namespace_gc_queue,
poll_period_seconds=POLL_PERIOD_SECONDS,
reservation_seconds=NAMESPACE_GC_TIMEOUT,
)
worker.start()