mirror of
https://github.com/facebook/proxygen.git
synced 2025-08-05 19:55:47 +03:00
getdeps: add build cache abstraction
Summary: This diff adds a small abstraction that allows for uploading and downloading from an artifact cache. We try to download from this cache at build time, but will only try to populate it for continuous builds--those are built from code that has been reviewed and landed on master. This restriction helps to avoid thrashing the cache with works in progress and results in a slightly more trustworthy state of the cache contents. In addition to this, we choose only to cache third party projects. The rationale is that our first party projects move too quickly to be worth caching, especially since the cache granularity is for the whole project rather than just changed elements of a given project. In a later diff I will introduce some implementations of the cache class that work with eg: Travis or Circle CI caching. Reviewed By: simpkins Differential Revision: D16873307 fbshipit-source-id: 2bfb69e36615791747b499073586562f2ca48be9
This commit is contained in:
committed by
Facebook Github Bot
parent
612c3d8e09
commit
2037c7f32a
@@ -14,7 +14,13 @@ import os
|
|||||||
import shutil
|
import shutil
|
||||||
import subprocess
|
import subprocess
|
||||||
import sys
|
import sys
|
||||||
|
import tarfile
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
# We don't import cache.create_cache directly as the facebook
|
||||||
|
# specific import below may monkey patch it, and we want to
|
||||||
|
# observe the patched version of this function!
|
||||||
|
import getdeps.cache as cache_module
|
||||||
from getdeps.buildopts import setup_build_options
|
from getdeps.buildopts import setup_build_options
|
||||||
from getdeps.dyndeps import create_dyn_dep_munger
|
from getdeps.dyndeps import create_dyn_dep_munger
|
||||||
from getdeps.errors import TransientFailure
|
from getdeps.errors import TransientFailure
|
||||||
@@ -157,6 +163,72 @@ class ProjectCmdBase(SubCmd):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class CachedProject(object):
|
||||||
|
""" A helper that allows calling the cache logic for a project
|
||||||
|
from both the build and the fetch code """
|
||||||
|
|
||||||
|
def __init__(self, cache, loader, m):
|
||||||
|
self.m = m
|
||||||
|
self.inst_dir = loader.get_project_install_dir(m)
|
||||||
|
self.project_hash = loader.get_project_hash(m)
|
||||||
|
self.ctx = loader.ctx_gen.get_context(m.name)
|
||||||
|
self.loader = loader
|
||||||
|
self.cache = cache
|
||||||
|
|
||||||
|
self.cache_file_name = "-".join(
|
||||||
|
(
|
||||||
|
m.name,
|
||||||
|
self.ctx.get("os"),
|
||||||
|
self.ctx.get("distro") or "none",
|
||||||
|
self.ctx.get("distro_vers") or "none",
|
||||||
|
self.project_hash,
|
||||||
|
"buildcache.tgz",
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
def is_cacheable(self):
|
||||||
|
""" We only cache third party projects """
|
||||||
|
return self.cache and not self.m.shipit_fbcode_builder
|
||||||
|
|
||||||
|
def download(self):
|
||||||
|
if self.is_cacheable() and not os.path.exists(self.inst_dir):
|
||||||
|
print("check cache for %s" % self.cache_file_name)
|
||||||
|
dl_dir = os.path.join(self.loader.build_opts.scratch_dir, "downloads")
|
||||||
|
if not os.path.exists(dl_dir):
|
||||||
|
os.makedirs(dl_dir)
|
||||||
|
try:
|
||||||
|
target_file_name = os.path.join(dl_dir, self.cache_file_name)
|
||||||
|
if self.cache.download_to_file(self.cache_file_name, target_file_name):
|
||||||
|
tf = tarfile.open(target_file_name, "r")
|
||||||
|
print(
|
||||||
|
"Extracting %s -> %s..." % (self.cache_file_name, self.inst_dir)
|
||||||
|
)
|
||||||
|
tf.extractall(self.inst_dir)
|
||||||
|
return True
|
||||||
|
except Exception as exc:
|
||||||
|
print("%s" % str(exc))
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def upload(self):
|
||||||
|
if self.cache and not self.m.shipit_fbcode_builder:
|
||||||
|
# We can prepare an archive and stick it in LFS
|
||||||
|
tempdir = tempfile.mkdtemp()
|
||||||
|
tarfilename = os.path.join(tempdir, self.cache_file_name)
|
||||||
|
print("Archiving for cache: %s..." % tarfilename)
|
||||||
|
tf = tarfile.open(tarfilename, "w:gz")
|
||||||
|
tf.add(self.inst_dir, arcname=".")
|
||||||
|
tf.close()
|
||||||
|
try:
|
||||||
|
self.cache.upload_from_file(self.cache_file_name, tarfilename)
|
||||||
|
except Exception as exc:
|
||||||
|
print(
|
||||||
|
"Failed to upload to cache (%s), continue anyway" % str(exc),
|
||||||
|
file=sys.stderr,
|
||||||
|
)
|
||||||
|
shutil.rmtree(tempdir)
|
||||||
|
|
||||||
|
|
||||||
@cmd("fetch", "fetch the code for a given project")
|
@cmd("fetch", "fetch the code for a given project")
|
||||||
class FetchCmd(ProjectCmdBase):
|
class FetchCmd(ProjectCmdBase):
|
||||||
def setup_project_cmd_parser(self, parser):
|
def setup_project_cmd_parser(self, parser):
|
||||||
@@ -179,7 +251,24 @@ class FetchCmd(ProjectCmdBase):
|
|||||||
projects = loader.manifests_in_dependency_order()
|
projects = loader.manifests_in_dependency_order()
|
||||||
else:
|
else:
|
||||||
projects = [manifest]
|
projects = [manifest]
|
||||||
|
|
||||||
|
cache = cache_module.create_cache()
|
||||||
for m in projects:
|
for m in projects:
|
||||||
|
cached_project = CachedProject(cache, loader, m)
|
||||||
|
if cached_project.download():
|
||||||
|
continue
|
||||||
|
|
||||||
|
inst_dir = loader.get_project_install_dir(m)
|
||||||
|
built_marker = os.path.join(inst_dir, ".built-by-getdeps")
|
||||||
|
if os.path.exists(built_marker):
|
||||||
|
with open(built_marker, "r") as f:
|
||||||
|
built_hash = f.read().strip()
|
||||||
|
|
||||||
|
project_hash = loader.get_project_hash(m)
|
||||||
|
if built_hash == project_hash:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# We need to fetch the sources
|
||||||
fetcher = loader.create_fetcher(m)
|
fetcher = loader.create_fetcher(m)
|
||||||
fetcher.update()
|
fetcher.update()
|
||||||
|
|
||||||
@@ -267,6 +356,8 @@ class BuildCmd(ProjectCmdBase):
|
|||||||
print("Building on %s" % loader.ctx_gen.get_context(args.project))
|
print("Building on %s" % loader.ctx_gen.get_context(args.project))
|
||||||
projects = loader.manifests_in_dependency_order()
|
projects = loader.manifests_in_dependency_order()
|
||||||
|
|
||||||
|
cache = cache_module.create_cache()
|
||||||
|
|
||||||
# Accumulate the install directories so that the build steps
|
# Accumulate the install directories so that the build steps
|
||||||
# can find their dep installation
|
# can find their dep installation
|
||||||
install_dirs = []
|
install_dirs = []
|
||||||
@@ -282,26 +373,20 @@ class BuildCmd(ProjectCmdBase):
|
|||||||
|
|
||||||
if m == manifest or not args.no_deps:
|
if m == manifest or not args.no_deps:
|
||||||
print("Assessing %s..." % m.name)
|
print("Assessing %s..." % m.name)
|
||||||
change_status = fetcher.update()
|
|
||||||
reconfigure = change_status.build_changed()
|
|
||||||
sources_changed = change_status.sources_changed()
|
|
||||||
|
|
||||||
project_hash = loader.get_project_hash(m)
|
project_hash = loader.get_project_hash(m)
|
||||||
|
ctx = loader.ctx_gen.get_context(m.name)
|
||||||
built_marker = os.path.join(inst_dir, ".built-by-getdeps")
|
built_marker = os.path.join(inst_dir, ".built-by-getdeps")
|
||||||
if os.path.exists(built_marker):
|
|
||||||
with open(built_marker, "r") as f:
|
cached_project = CachedProject(cache, loader, m)
|
||||||
built_hash = f.read().strip()
|
|
||||||
if built_hash != project_hash:
|
reconfigure, sources_changed = self.compute_source_change_status(
|
||||||
# Some kind of inconsistency with a prior build,
|
cached_project, fetcher, m, built_marker, project_hash
|
||||||
# let's run it again to be sure
|
)
|
||||||
os.unlink(built_marker)
|
|
||||||
reconfigure = True
|
|
||||||
|
|
||||||
if sources_changed or reconfigure or not os.path.exists(built_marker):
|
if sources_changed or reconfigure or not os.path.exists(built_marker):
|
||||||
if os.path.exists(built_marker):
|
if os.path.exists(built_marker):
|
||||||
os.unlink(built_marker)
|
os.unlink(built_marker)
|
||||||
src_dir = fetcher.get_src_dir()
|
src_dir = fetcher.get_src_dir()
|
||||||
ctx = loader.ctx_gen.get_context(m.name)
|
|
||||||
builder = m.create_builder(
|
builder = m.create_builder(
|
||||||
loader.build_opts, src_dir, build_dir, inst_dir, ctx
|
loader.build_opts, src_dir, build_dir, inst_dir, ctx
|
||||||
)
|
)
|
||||||
@@ -310,8 +395,46 @@ class BuildCmd(ProjectCmdBase):
|
|||||||
with open(built_marker, "w") as f:
|
with open(built_marker, "w") as f:
|
||||||
f.write(project_hash)
|
f.write(project_hash)
|
||||||
|
|
||||||
|
# Only populate the cache from continuous build runs
|
||||||
|
if args.schedule_type == "continuous":
|
||||||
|
cached_project.upload()
|
||||||
|
|
||||||
install_dirs.append(inst_dir)
|
install_dirs.append(inst_dir)
|
||||||
|
|
||||||
|
def compute_source_change_status(
|
||||||
|
self, cached_project, fetcher, m, built_marker, project_hash
|
||||||
|
):
|
||||||
|
reconfigure = False
|
||||||
|
sources_changed = False
|
||||||
|
if not cached_project.download():
|
||||||
|
check_fetcher = True
|
||||||
|
if os.path.exists(built_marker):
|
||||||
|
check_fetcher = False
|
||||||
|
with open(built_marker, "r") as f:
|
||||||
|
built_hash = f.read().strip()
|
||||||
|
if built_hash == project_hash:
|
||||||
|
if cached_project.is_cacheable():
|
||||||
|
# We can blindly trust the build status
|
||||||
|
reconfigure = False
|
||||||
|
sources_changed = False
|
||||||
|
else:
|
||||||
|
# Otherwise, we may have changed the source, so let's
|
||||||
|
# check in with the fetcher layer
|
||||||
|
check_fetcher = True
|
||||||
|
else:
|
||||||
|
# Some kind of inconsistency with a prior build,
|
||||||
|
# let's run it again to be sure
|
||||||
|
os.unlink(built_marker)
|
||||||
|
reconfigure = True
|
||||||
|
sources_changed = True
|
||||||
|
|
||||||
|
if check_fetcher:
|
||||||
|
change_status = fetcher.update()
|
||||||
|
reconfigure = change_status.build_changed()
|
||||||
|
sources_changed = change_status.sources_changed()
|
||||||
|
|
||||||
|
return reconfigure, sources_changed
|
||||||
|
|
||||||
def setup_project_cmd_parser(self, parser):
|
def setup_project_cmd_parser(self, parser):
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--clean",
|
"--clean",
|
||||||
@@ -333,6 +456,9 @@ class BuildCmd(ProjectCmdBase):
|
|||||||
"slow up-to-date-ness checks"
|
"slow up-to-date-ness checks"
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"--schedule-type", help="Indicates how the build was activated"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@cmd("fixup-dyn-deps", "Adjusts dynamic dependencies for packaging purposes")
|
@cmd("fixup-dyn-deps", "Adjusts dynamic dependencies for packaging purposes")
|
||||||
|
41
build/fbcode_builder/getdeps/cache.py
Normal file
41
build/fbcode_builder/getdeps/cache.py
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
# Copyright (c) 2019-present, Facebook, Inc.
|
||||||
|
# All rights reserved.
|
||||||
|
#
|
||||||
|
# This source code is licensed under the BSD-style license found in the
|
||||||
|
# LICENSE file in the root directory of this source tree. An additional grant
|
||||||
|
# of patent rights can be found in the PATENTS file in the same directory.
|
||||||
|
|
||||||
|
from __future__ import absolute_import, division, print_function, unicode_literals
|
||||||
|
|
||||||
|
|
||||||
|
class ArtifactCache(object):
|
||||||
|
""" The ArtifactCache is a small abstraction that allows caching
|
||||||
|
named things in some external storage mechanism.
|
||||||
|
The primary use case is for storing the build products on CI
|
||||||
|
systems to accelerate the build """
|
||||||
|
|
||||||
|
def download_to_file(self, name, dest_file_name):
|
||||||
|
""" If `name` exists in the cache, download it and place it
|
||||||
|
in the specified `dest_file_name` location on the filesystem.
|
||||||
|
If a transient issue was encountered a TransientFailure shall
|
||||||
|
be raised.
|
||||||
|
If `name` doesn't exist in the cache `False` shall be returned.
|
||||||
|
If `dest_file_name` was successfully updated `True` shall be
|
||||||
|
returned.
|
||||||
|
All other conditions shall raise an appropriate exception. """
|
||||||
|
return False
|
||||||
|
|
||||||
|
def upload_from_file(self, name, source_file_name):
|
||||||
|
""" Causes `name` to be populated in the cache by uploading
|
||||||
|
the contents of `source_file_name` to the storage system.
|
||||||
|
If a transient issue was encountered a TransientFailure shall
|
||||||
|
be raised.
|
||||||
|
If the upload failed for some other reason, an appropriate
|
||||||
|
exception shall be raised. """
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def create_cache():
|
||||||
|
""" This function is monkey patchable to provide an actual
|
||||||
|
implementation """
|
||||||
|
return None
|
Reference in New Issue
Block a user