Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 #!/usr/bin/env python | 1 #!/usr/bin/env python |
| 2 # Copyright 2014 The Chromium Authors. All rights reserved. | 2 # Copyright 2014 The Chromium Authors. All rights reserved. |
| 3 # Use of this source code is governed by a BSD-style license that can be | 3 # Use of this source code is governed by a BSD-style license that can be |
| 4 # found in the LICENSE file. | 4 # found in the LICENSE file. |
| 5 | 5 |
| 6 """A git command for managing a local cache of git repositories.""" | 6 """A git command for managing a local cache of git repositories.""" |
| 7 | 7 |
| 8 from __future__ import print_function | |
| 8 import errno | 9 import errno |
| 9 import logging | 10 import logging |
| 10 import optparse | 11 import optparse |
| 11 import os | 12 import os |
| 12 import tempfile | 13 import tempfile |
| 13 import subprocess | 14 import subprocess |
| 14 import sys | 15 import sys |
| 15 import urlparse | 16 import urlparse |
| 16 | 17 |
| 17 from download_from_google_storage import Gsutil | 18 from download_from_google_storage import Gsutil |
| 18 import gclient_utils | 19 import gclient_utils |
| 19 import subcommand | 20 import subcommand |
| 20 | 21 |
| 21 | 22 try: |
| 22 GIT_EXECUTABLE = 'git.bat' if sys.platform.startswith('win') else 'git' | 23 # pylint: disable=E0602 |
| 23 BOOTSTRAP_BUCKET = 'chromium-git-cache' | 24 WinErr = WindowsError |
| 24 GSUTIL_DEFAULT_PATH = os.path.join( | 25 except NameError: |
| 25 os.path.dirname(os.path.abspath(__file__)), | 26 class WinErr(Exception): |
| 26 'third_party', 'gsutil', 'gsutil') | 27 pass |
| 27 | |
| 28 | |
| 29 def UrlToCacheDir(url): | |
| 30 """Convert a git url to a normalized form for the cache dir path.""" | |
| 31 parsed = urlparse.urlparse(url) | |
| 32 norm_url = parsed.netloc + parsed.path | |
| 33 if norm_url.endswith('.git'): | |
| 34 norm_url = norm_url[:-len('.git')] | |
| 35 return norm_url.replace('-', '--').replace('/', '-').lower() | |
| 36 | |
| 37 | |
| 38 def RunGit(cmd, **kwargs): | |
| 39 """Run git in a subprocess.""" | |
| 40 kwargs.setdefault('cwd', os.getcwd()) | |
| 41 if kwargs.get('filter_fn'): | |
| 42 kwargs['filter_fn'] = gclient_utils.GitFilter(kwargs.get('filter_fn')) | |
| 43 kwargs.setdefault('print_stdout', False) | |
| 44 env = kwargs.get('env') or kwargs.setdefault('env', os.environ.copy()) | |
| 45 env.setdefault('GIT_ASKPASS', 'true') | |
| 46 env.setdefault('SSH_ASKPASS', 'true') | |
| 47 else: | |
| 48 kwargs.setdefault('print_stdout', True) | |
| 49 stdout = kwargs.get('stdout', sys.stdout) | |
| 50 print >> stdout, 'running "git %s" in "%s"' % (' '.join(cmd), kwargs['cwd']) | |
| 51 gclient_utils.CheckCallAndFilter([GIT_EXECUTABLE] + cmd, **kwargs) | |
| 52 | |
| 53 | 28 |
| 54 class LockError(Exception): | 29 class LockError(Exception): |
| 55 pass | 30 pass |
| 56 | 31 |
| 57 | 32 |
| 58 class Lockfile(object): | 33 class Lockfile(object): |
| 59 """Class to represent a cross-platform process-specific lockfile.""" | 34 """Class to represent a cross-platform process-specific lockfile.""" |
| 60 | 35 |
| 61 def __init__(self, path): | 36 def __init__(self, path): |
| 62 self.path = os.path.abspath(path) | 37 self.path = os.path.abspath(path) |
| (...skipping 11 matching lines...) Expand all Loading... | |
| 74 pid = int(f.readline().strip()) | 49 pid = int(f.readline().strip()) |
| 75 except (IOError, ValueError): | 50 except (IOError, ValueError): |
| 76 pid = None | 51 pid = None |
| 77 return pid | 52 return pid |
| 78 | 53 |
| 79 def _make_lockfile(self): | 54 def _make_lockfile(self): |
| 80 """Safely creates a lockfile containing the current pid.""" | 55 """Safely creates a lockfile containing the current pid.""" |
| 81 open_flags = (os.O_CREAT | os.O_EXCL | os.O_WRONLY) | 56 open_flags = (os.O_CREAT | os.O_EXCL | os.O_WRONLY) |
| 82 fd = os.open(self.lockfile, open_flags, 0o644) | 57 fd = os.open(self.lockfile, open_flags, 0o644) |
| 83 f = os.fdopen(fd, 'w') | 58 f = os.fdopen(fd, 'w') |
| 84 print >> f, self.pid | 59 print(self.pid, file=f) |
| 85 f.close() | 60 f.close() |
| 86 | 61 |
| 87 def _remove_lockfile(self): | 62 def _remove_lockfile(self): |
| 88 """Delete the lockfile. Complains (implicitly) if it doesn't exist.""" | 63 """Delete the lockfile. Complains (implicitly) if it doesn't exist.""" |
| 89 os.remove(self.lockfile) | 64 os.remove(self.lockfile) |
| 90 | 65 |
| 91 def lock(self): | 66 def lock(self): |
| 92 """Acquire the lock. | 67 """Acquire the lock. |
| 93 | 68 |
| 94 Note: This is a NON-BLOCKING FAIL-FAST operation. | 69 Note: This is a NON-BLOCKING FAIL-FAST operation. |
| (...skipping 36 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 131 | 106 |
| 132 def i_am_locking(self): | 107 def i_am_locking(self): |
| 133 """Test if the file is locked by this process.""" | 108 """Test if the file is locked by this process.""" |
| 134 return self.is_locked() and self.pid == self._read_pid() | 109 return self.is_locked() and self.pid == self._read_pid() |
| 135 | 110 |
| 136 def __enter__(self): | 111 def __enter__(self): |
| 137 self.lock() | 112 self.lock() |
| 138 return self | 113 return self |
| 139 | 114 |
| 140 def __exit__(self, *_exc): | 115 def __exit__(self, *_exc): |
| 141 self.unlock() | 116 # Windows is unreliable when it comes to file locking. YMMV. |
| 142 | 117 try: |
| 118 self.unlock() | |
| 119 except WinErr: | |
| 120 pass | |
|
Ryan Tseng
2014/04/09 00:39:44
if os.exists(self.lockfile):
raise
If the lock
szager1
2014/04/09 05:22:29
Below is sporadic error I see on Windows. There's
| |
| 121 | |
| 122 | |
| 123 class Mirror(object): | |
| 124 | |
| 125 git_exe = 'git.bat' if sys.platform.startswith('win') else 'git' | |
| 126 gsutil_exe = os.path.join( | |
| 127 os.path.dirname(os.path.abspath(__file__)), | |
| 128 'third_party', 'gsutil', 'gsutil') | |
| 129 bootstrap_bucket = 'chromium-git-cache' | |
| 130 | |
| 131 def __init__(self, url, refs=None, print_func=None): | |
| 132 self.url = url | |
| 133 self.refs = refs or [] | |
| 134 self.basedir = self.UrlToCacheDir(url) | |
| 135 self.mirror_path = os.path.join(self.GetCachePath(), self.basedir) | |
| 136 self.print = print_func or print | |
| 137 | |
| 138 @staticmethod | |
| 139 def UrlToCacheDir(url): | |
| 140 """Convert a git url to a normalized form for the cache dir path.""" | |
| 141 parsed = urlparse.urlparse(url) | |
| 142 norm_url = parsed.netloc + parsed.path | |
| 143 if norm_url.endswith('.git'): | |
| 144 norm_url = norm_url[:-len('.git')] | |
| 145 return norm_url.replace('-', '--').replace('/', '-').lower() | |
| 146 | |
| 147 @staticmethod | |
| 148 def FindExecutable(executable): | |
|
agable
2014/04/11 20:44:58
Why is this a method on Mirror?
| |
| 149 """This mimics the "which" utility.""" | |
| 150 path_folders = os.environ.get('PATH').split(os.pathsep) | |
| 151 | |
| 152 for path_folder in path_folders: | |
| 153 target = os.path.join(path_folder, executable) | |
| 154 # Just incase we have some ~/blah paths. | |
| 155 target = os.path.abspath(os.path.expanduser(target)) | |
| 156 if os.path.isfile(target) and os.access(target, os.X_OK): | |
| 157 return target | |
| 158 return None | |
| 159 | |
| 160 @classmethod | |
| 161 def SetCachePath(cls, cachepath): | |
|
agable
2014/04/11 20:44:58
Could make cache_path a property with a @property
| |
| 162 setattr(cls, 'cachepath', cachepath) | |
| 163 | |
| 164 @classmethod | |
| 165 def GetCachePath(cls): | |
|
Ryan Tseng
2014/04/09 00:39:44
This is rather convoluted, and I feel like it'll c
szager1
2014/04/09 05:22:29
Unlike url, cachepath is a global setting. It doe
agable
2014/04/11 20:44:58
Can't do this at __init__ time because it's a clas
| |
| 166 if not hasattr(cls, 'cachepath'): | |
| 167 try: | |
| 168 cachepath = subprocess.check_output( | |
| 169 [cls.git_exe, 'config', '--global', 'cache.cachepath']).strip() | |
| 170 except subprocess.CalledProcessError: | |
| 171 cachepath = None | |
| 172 if not cachepath: | |
| 173 raise RuntimeError('No global cache.cachepath git configuration found.') | |
| 174 setattr(cls, 'cachepath', cachepath) | |
| 175 return getattr(cls, 'cachepath') | |
| 176 | |
| 177 def RunGit(self, cmd, **kwargs): | |
| 178 """Run git in a subprocess.""" | |
| 179 cwd = kwargs.setdefault('cwd', self.mirror_path) | |
| 180 kwargs.setdefault('print_stdout', False) | |
| 181 kwargs.setdefault('filter_fn', self.print) | |
| 182 env = kwargs.get('env') or kwargs.setdefault('env', os.environ.copy()) | |
| 183 env.setdefault('GIT_ASKPASS', 'true') | |
| 184 env.setdefault('SSH_ASKPASS', 'true') | |
| 185 self.print('running "git %s" in "%s"' % (' '.join(cmd), cwd)) | |
| 186 gclient_utils.CheckCallAndFilter([self.git_exe] + cmd, **kwargs) | |
| 187 | |
| 188 def config(self, cwd=None): | |
| 189 if cwd is None: | |
| 190 cwd = self.mirror_path | |
| 191 self.RunGit(['config', 'core.deltaBaseCacheLimit', | |
| 192 gclient_utils.DefaultDeltaBaseCacheLimit()], cwd=cwd) | |
| 193 self.RunGit(['config', 'remote.origin.url', self.url], cwd=cwd) | |
| 194 self.RunGit(['config', '--replace-all', 'remote.origin.fetch', | |
| 195 '+refs/heads/*:refs/heads/*'], cwd=cwd) | |
| 196 for ref in self.refs: | |
| 197 ref = ref.lstrip('+').rstrip('/') | |
| 198 if ref.startswith('refs/'): | |
| 199 refspec = '+%s:%s' % (ref, ref) | |
| 200 else: | |
| 201 refspec = '+refs/%s/*:refs/%s/*' % (ref, ref) | |
| 202 self.RunGit(['config', '--add', 'remote.origin.fetch', refspec], cwd=cwd) | |
| 203 | |
| 204 def bootstrap_repo(self, directory): | |
| 205 """Bootstrap the repo from Google Stroage if possible. | |
| 206 | |
| 207 Requires 7z on Windows and Unzip on Linux/Mac. | |
| 208 """ | |
| 209 if sys.platform.startswith('win'): | |
| 210 if not self.FindExecutable('7z'): | |
| 211 self.print(''' | |
|
agable
2014/04/11 20:44:58
use textwrap.dedent.
| |
| 212 Cannot find 7z in the path. If you want git cache to be able to bootstrap from | |
|
Ryan Tseng
2014/04/09 00:39:44
nit: 2 spaces here, or 1 space down on line 221/11
szager1
2014/04/09 05:22:29
Done.
| |
| 213 Google Storage, please install 7z from: | |
| 214 | |
| 215 http://www.7-zip.org/download.html | |
| 216 ''') | |
| 217 return False | |
| 218 else: | |
| 219 if not self.FindExecutable('unzip'): | |
| 220 self.print(''' | |
| 221 Cannot find unzip in the path. If you want git cache to be able to bootstrap | |
| 222 from Google Storage, please ensure unzip is present on your system. | |
| 223 ''') | |
| 224 return False | |
| 225 | |
| 226 gs_folder = 'gs://%s/%s' % (self.bootstrap_bucket, self.basedir) | |
| 227 gsutil = Gsutil( | |
| 228 self.gsutil_exe, boto_path=os.devnull, bypass_prodaccess=True) | |
| 229 # Get the most recent version of the zipfile. | |
| 230 _, ls_out, _ = gsutil.check_call('ls', gs_folder) | |
| 231 ls_out_sorted = sorted(ls_out.splitlines()) | |
| 232 if not ls_out_sorted: | |
| 233 # This repo is not on Google Storage. | |
| 234 return False | |
| 235 latest_checkout = ls_out_sorted[-1] | |
| 236 | |
| 237 # Download zip file to a temporary directory. | |
| 238 try: | |
| 239 tempdir = tempfile.mkdtemp() | |
| 240 self.print('Downloading %s' % latest_checkout) | |
| 241 code, out, err = gsutil.check_call('cp', latest_checkout, tempdir) | |
| 242 if code: | |
| 243 self.print('%s\n%s' % (out, err)) | |
| 244 return False | |
| 245 filename = os.path.join(tempdir, latest_checkout.split('/')[-1]) | |
| 246 | |
| 247 # Unpack the file with 7z on Windows, or unzip everywhere else. | |
| 248 if sys.platform.startswith('win'): | |
| 249 cmd = ['7z', 'x', '-o%s' % directory, '-tzip', filename] | |
| 250 else: | |
| 251 cmd = ['unzip', filename, '-d', directory] | |
| 252 retcode = subprocess.call(cmd) | |
| 253 finally: | |
| 254 # Clean up the downloaded zipfile. | |
| 255 gclient_utils.rmtree(tempdir) | |
| 256 | |
| 257 if retcode: | |
| 258 self.print( | |
| 259 'Extracting bootstrap zipfile %s failed.\n' | |
| 260 'Resuming normal operations.' % filename) | |
| 261 return False | |
| 262 return True | |
| 263 | |
| 264 def exists(self): | |
| 265 return os.path.isfile(os.path.join(self.mirror_path, 'config')) | |
| 266 | |
| 267 def populate(self, depth=None, shallow=False, bootstrap=False, noisy=False): | |
| 268 if shallow and not depth: | |
| 269 depth = 10000 | |
| 270 gclient_utils.safe_makedirs(self.GetCachePath()) | |
| 271 | |
| 272 v = [] | |
| 273 if noisy: | |
|
Ryan Tseng
2014/04/09 00:39:44
why not just call it verbose all the way down the
szager1
2014/04/09 05:22:29
Done.
| |
| 274 v = ['-v', '--progress'] | |
| 275 | |
| 276 d = [] | |
| 277 if depth: | |
| 278 d = ['--depth', str(depth)] | |
| 279 | |
| 280 | |
| 281 with Lockfile(self.mirror_path): | |
| 282 # Setup from scratch if the repo is new or is in a bad state. | |
| 283 tempdir = None | |
| 284 if not os.path.exists(os.path.join(self.mirror_path, 'config')): | |
| 285 gclient_utils.rmtree(self.mirror_path) | |
| 286 tempdir = tempfile.mkdtemp( | |
| 287 suffix=self.basedir, dir=self.GetCachePath()) | |
| 288 bootstrapped = bootstrap and self.bootstrap_repo(tempdir) | |
|
Ryan Tseng
2014/04/09 00:39:44
Actually we probably want
not depth and bootstrap
szager1
2014/04/09 05:22:29
Done.
| |
| 289 if not bootstrapped: | |
| 290 self.RunGit(['init', '--bare'], cwd=tempdir) | |
| 291 else: | |
| 292 if depth and os.path.exists(os.path.join(self.mirror_path, 'shallow')): | |
| 293 logging.warn( | |
| 294 'Shallow fetch requested, but repo cache already exists.') | |
| 295 d = [] | |
| 296 | |
| 297 rundir = tempdir or self.mirror_path | |
| 298 self.config(rundir) | |
| 299 fetch_cmd = ['fetch'] + v + d + ['origin'] | |
| 300 fetch_specs = subprocess.check_output( | |
| 301 [self.git_exe, 'config', '--get-all', 'remote.origin.fetch'], | |
| 302 cwd=rundir).strip().splitlines() | |
| 303 for spec in fetch_specs: | |
| 304 try: | |
| 305 self.RunGit(fetch_cmd + [spec], cwd=rundir, retry=True) | |
| 306 except subprocess.CalledProcessError: | |
| 307 logging.warn('Fetch of %s failed' % spec) | |
| 308 if tempdir: | |
| 309 os.rename(tempdir, self.mirror_path) | |
| 310 | |
| 311 def update_bootstrap(self): | |
| 312 # The files are named <git number>.zip | |
| 313 gen_number = subprocess.check_output( | |
| 314 [self.git_exe, 'number', 'master'], cwd=self.mirror_path).strip() | |
| 315 self.RunGit(['gc']) # Run Garbage Collect to compress packfile. | |
| 316 # Creating a temp file and then deleting it ensures we can use this name. | |
| 317 _, tmp_zipfile = tempfile.mkstemp(suffix='.zip') | |
| 318 os.remove(tmp_zipfile) | |
| 319 subprocess.call(['zip', '-r', tmp_zipfile, '.'], cwd=self.mirror_path) | |
| 320 gsutil = Gsutil(path=self.gsutil_exe, boto_path=None) | |
| 321 dest_name = 'gs://%s/%s/%s.zip' % ( | |
| 322 self.bootstrap_bucket, self.basedir, gen_number) | |
| 323 gsutil.call('cp', tmp_zipfile, dest_name) | |
| 324 os.remove(tmp_zipfile) | |
| 325 | |
| 326 def unlock(self): | |
| 327 lf = Lockfile(self.mirror_path) | |
| 328 config_lock = os.path.join(self.mirror_path, 'config.lock') | |
| 329 if os.path.exists(config_lock): | |
| 330 os.remove(config_lock) | |
| 331 lf.break_lock() | |
| 143 | 332 |
| 144 @subcommand.usage('[url of repo to check for caching]') | 333 @subcommand.usage('[url of repo to check for caching]') |
| 145 def CMDexists(parser, args): | 334 def CMDexists(parser, args): |
| 146 """Check to see if there already is a cache of the given repo.""" | 335 """Check to see if there already is a cache of the given repo.""" |
| 147 options, args = parser.parse_args(args) | 336 _, args = parser.parse_args(args) |
| 148 if not len(args) == 1: | 337 if not len(args) == 1: |
| 149 parser.error('git cache exists only takes exactly one repo url.') | 338 parser.error('git cache exists only takes exactly one repo url.') |
| 150 url = args[0] | 339 url = args[0] |
| 151 repo_dir = os.path.join(options.cache_dir, UrlToCacheDir(url)) | 340 mirror = Mirror(url) |
| 152 flag_file = os.path.join(repo_dir, 'config') | 341 if mirror.exists(): |
| 153 if os.path.isdir(repo_dir) and os.path.isfile(flag_file): | 342 print(mirror.mirror_path) |
| 154 print repo_dir | |
| 155 return 0 | 343 return 0 |
| 156 return 1 | 344 return 1 |
| 157 | 345 |
| 158 | 346 |
| 159 @subcommand.usage('[url of repo to create a bootstrap zip file]') | 347 @subcommand.usage('[url of repo to create a bootstrap zip file]') |
| 160 def CMDupdate_bootstrap(parser, args): | 348 def CMDupdate_bootstrap(parser, args): |
| 161 """Create and uploads a bootstrap tarball.""" | 349 """Create and uploads a bootstrap tarball.""" |
| 162 # Lets just assert we can't do this on Windows. | 350 # Lets just assert we can't do this on Windows. |
| 163 if sys.platform.startswith('win'): | 351 if sys.platform.startswith('win'): |
| 164 print >> sys.stderr, 'Sorry, update bootstrap will not work on Windows.' | 352 print('Sorry, update bootstrap will not work on Windows.', file=sys.stderr) |
| 165 return 1 | 353 return 1 |
| 166 | 354 |
| 167 # First, we need to ensure the cache is populated. | 355 # First, we need to ensure the cache is populated. |
| 168 populate_args = args[:] | 356 populate_args = args[:] |
| 169 populate_args.append('--no_bootstrap') | 357 populate_args.append('--no_bootstrap') |
| 170 CMDpopulate(parser, populate_args) | 358 CMDpopulate(parser, populate_args) |
| 171 | 359 |
| 172 # Get the repo directory. | 360 # Get the repo directory. |
| 173 options, args = parser.parse_args(args) | 361 _, args = parser.parse_args(args) |
| 174 url = args[0] | 362 url = args[0] |
| 175 repo_dir = os.path.join(options.cache_dir, UrlToCacheDir(url)) | 363 mirror = Mirror(url) |
| 176 | 364 mirror.update_bootstrap() |
| 177 # The files are named <git number>.zip | 365 return 0 |
| 178 gen_number = subprocess.check_output(['git', 'number', 'master'], | |
| 179 cwd=repo_dir).strip() | |
| 180 RunGit(['gc'], cwd=repo_dir) # Run Garbage Collect to compress packfile. | |
| 181 # Creating a temp file and then deleting it ensures we can use this name. | |
| 182 _, tmp_zipfile = tempfile.mkstemp(suffix='.zip') | |
| 183 os.remove(tmp_zipfile) | |
| 184 subprocess.call(['zip', '-r', tmp_zipfile, '.'], cwd=repo_dir) | |
| 185 gsutil = Gsutil(path=GSUTIL_DEFAULT_PATH, boto_path=None) | |
| 186 dest_name = 'gs://%s/%s/%s.zip' % (BOOTSTRAP_BUCKET, | |
| 187 UrlToCacheDir(url), | |
| 188 gen_number) | |
| 189 gsutil.call('cp', tmp_zipfile, dest_name) | |
| 190 os.remove(tmp_zipfile) | |
| 191 | 366 |
| 192 | 367 |
| 193 @subcommand.usage('[url of repo to add to or update in cache]') | 368 @subcommand.usage('[url of repo to add to or update in cache]') |
| 194 def CMDpopulate(parser, args): | 369 def CMDpopulate(parser, args): |
| 195 """Ensure that the cache has all up-to-date objects for the given repo.""" | 370 """Ensure that the cache has all up-to-date objects for the given repo.""" |
| 196 parser.add_option('--depth', type='int', | 371 parser.add_option('--depth', type='int', |
| 197 help='Only cache DEPTH commits of history') | 372 help='Only cache DEPTH commits of history') |
| 198 parser.add_option('--shallow', '-s', action='store_true', | 373 parser.add_option('--shallow', '-s', action='store_true', |
| 199 help='Only cache 10000 commits of history') | 374 help='Only cache 10000 commits of history') |
| 200 parser.add_option('--ref', action='append', | 375 parser.add_option('--ref', action='append', |
| 201 help='Specify additional refs to be fetched') | 376 help='Specify additional refs to be fetched') |
| 202 parser.add_option('--no_bootstrap', action='store_true', | 377 parser.add_option('--no_bootstrap', action='store_true', |
| 203 help='Don\'t bootstrap from Google Storage') | 378 help='Don\'t bootstrap from Google Storage') |
| 204 | 379 |
| 205 options, args = parser.parse_args(args) | 380 options, args = parser.parse_args(args) |
| 206 if options.shallow and not options.depth: | 381 if options.shallow and not options.depth: |
| 207 options.depth = 10000 | 382 options.depth = 10000 |
|
Ryan Tseng
2014/04/09 00:39:44
no longer needed.
szager1
2014/04/09 05:22:29
Done.
| |
| 208 if not len(args) == 1: | 383 if not len(args) == 1: |
| 209 parser.error('git cache populate only takes exactly one repo url.') | 384 parser.error('git cache populate only takes exactly one repo url.') |
| 210 url = args[0] | 385 url = args[0] |
| 211 | 386 |
| 212 gclient_utils.safe_makedirs(options.cache_dir) | 387 mirror = Mirror(url, refs=options.ref) |
| 213 repo_dir = os.path.join(options.cache_dir, UrlToCacheDir(url)) | 388 kwargs = { |
| 214 | 389 'noisy': options.verbose, |
| 215 v = [] | 390 'shallow': options.shallow, |
| 216 filter_fn = lambda l: '[up to date]' not in l | 391 'bootstrap': not options.no_bootstrap, |
| 217 if options.verbose: | 392 } |
| 218 v = ['-v', '--progress'] | |
| 219 filter_fn = None | |
| 220 | |
| 221 d = [] | |
| 222 if options.depth: | 393 if options.depth: |
| 223 d = ['--depth', '%d' % options.depth] | 394 kwargs['depth'] = options.depth |
| 224 | 395 mirror.populate(**kwargs) |
| 225 def _find(executable): | |
| 226 """This mimics the "which" utility.""" | |
| 227 path_folders = os.environ.get('PATH').split(os.pathsep) | |
| 228 | |
| 229 for path_folder in path_folders: | |
| 230 target = os.path.join(path_folder, executable) | |
| 231 # Just incase we have some ~/blah paths. | |
| 232 target = os.path.abspath(os.path.expanduser(target)) | |
| 233 if os.path.isfile(target) and os.access(target, os.X_OK): | |
| 234 return target | |
| 235 return False | |
| 236 | |
| 237 def _maybe_bootstrap_repo(directory): | |
| 238 """Bootstrap the repo from Google Stroage if possible. | |
| 239 | |
| 240 Requires 7z on Windows and Unzip on Linux/Mac. | |
| 241 """ | |
| 242 if options.no_bootstrap: | |
| 243 return False | |
| 244 if sys.platform.startswith('win'): | |
| 245 if not _find('7z'): | |
| 246 print 'Cannot find 7z in the path.' | |
| 247 print 'If you want git cache to be able to bootstrap from ' | |
| 248 print 'Google Storage, please install 7z from:' | |
| 249 print 'http://www.7-zip.org/download.html' | |
| 250 return False | |
| 251 else: | |
| 252 if not _find('unzip'): | |
| 253 print 'Cannot find unzip in the path.' | |
| 254 print 'If you want git cache to be able to bootstrap from ' | |
| 255 print 'Google Storage, please ensure unzip is present on your system.' | |
| 256 return False | |
| 257 | |
| 258 folder = UrlToCacheDir(url) | |
| 259 gs_folder = 'gs://%s/%s' % (BOOTSTRAP_BUCKET, folder) | |
| 260 gsutil = Gsutil(GSUTIL_DEFAULT_PATH, boto_path=os.devnull, | |
| 261 bypass_prodaccess=True) | |
| 262 # Get the most recent version of the zipfile. | |
| 263 _, ls_out, _ = gsutil.check_call('ls', gs_folder) | |
| 264 ls_out_sorted = sorted(ls_out.splitlines()) | |
| 265 if not ls_out_sorted: | |
| 266 # This repo is not on Google Storage. | |
| 267 return False | |
| 268 latest_checkout = ls_out_sorted[-1] | |
| 269 | |
| 270 # Download zip file to a temporary directory. | |
| 271 tempdir = tempfile.mkdtemp() | |
| 272 print 'Downloading %s...' % latest_checkout | |
| 273 code, out, err = gsutil.check_call('cp', latest_checkout, tempdir) | |
| 274 if code: | |
| 275 print '%s\n%s' % (out, err) | |
| 276 return False | |
| 277 filename = os.path.join(tempdir, latest_checkout.split('/')[-1]) | |
| 278 | |
| 279 # Unpack the file with 7z on Windows, or unzip everywhere else. | |
| 280 if sys.platform.startswith('win'): | |
| 281 cmd = ['7z', 'x', '-o%s' % directory, '-tzip', filename] | |
| 282 else: | |
| 283 cmd = ['unzip', filename, '-d', directory] | |
| 284 retcode = subprocess.call(cmd) | |
| 285 | |
| 286 # Clean up the downloaded zipfile. | |
| 287 gclient_utils.rmtree(tempdir) | |
| 288 if retcode: | |
| 289 print 'Extracting bootstrap zipfile %s failed.' % filename | |
| 290 print 'Resuming normal operations' | |
| 291 return False | |
| 292 return True | |
| 293 | |
| 294 def _config(directory): | |
| 295 RunGit(['config', 'core.deltaBaseCacheLimit', | |
| 296 gclient_utils.DefaultDeltaBaseCacheLimit()], cwd=directory) | |
| 297 RunGit(['config', 'remote.origin.url', url], | |
| 298 cwd=directory) | |
| 299 RunGit(['config', '--replace-all', 'remote.origin.fetch', | |
| 300 '+refs/heads/*:refs/heads/*'], | |
| 301 cwd=directory) | |
| 302 RunGit(['config', '--add', 'remote.origin.fetch', | |
| 303 '+refs/tags/*:refs/tags/*'], | |
| 304 cwd=directory) | |
| 305 for ref in options.ref or []: | |
| 306 ref = ref.rstrip('/') | |
| 307 refspec = '+refs/%s/*:refs/%s/*' % (ref, ref) | |
| 308 RunGit(['config', '--add', 'remote.origin.fetch', refspec], | |
| 309 cwd=directory) | |
| 310 | |
| 311 with Lockfile(repo_dir): | |
| 312 # Setup from scratch if the repo is new or is in a bad state. | |
| 313 if not os.path.exists(os.path.join(repo_dir, 'config')): | |
| 314 gclient_utils.rmtree(repo_dir) | |
| 315 tempdir = tempfile.mkdtemp(suffix=UrlToCacheDir(url), | |
| 316 dir=options.cache_dir) | |
| 317 bootstrapped = _maybe_bootstrap_repo(tempdir) | |
| 318 if not bootstrapped: | |
| 319 RunGit(['init', '--bare'], cwd=tempdir) | |
| 320 _config(tempdir) | |
| 321 fetch_cmd = ['fetch'] + v + d + ['origin'] | |
| 322 RunGit(fetch_cmd, filter_fn=filter_fn, cwd=tempdir, retry=True) | |
| 323 os.rename(tempdir, repo_dir) | |
| 324 else: | |
| 325 _config(repo_dir) | |
| 326 if options.depth and os.path.exists(os.path.join(repo_dir, 'shallow')): | |
| 327 logging.warn('Shallow fetch requested, but repo cache already exists.') | |
| 328 fetch_cmd = ['fetch'] + v + ['origin'] | |
| 329 RunGit(fetch_cmd, filter_fn=filter_fn, cwd=repo_dir, retry=True) | |
| 330 | 396 |
| 331 | 397 |
| 332 @subcommand.usage('[url of repo to unlock, or -a|--all]') | 398 @subcommand.usage('[url of repo to unlock, or -a|--all]') |
| 333 def CMDunlock(parser, args): | 399 def CMDunlock(parser, args): |
| 334 """Unlock one or all repos if their lock files are still around.""" | 400 """Unlock one or all repos if their lock files are still around.""" |
| 335 parser.add_option('--force', '-f', action='store_true', | 401 parser.add_option('--force', '-f', action='store_true', |
| 336 help='Actually perform the action') | 402 help='Actually perform the action') |
| 337 parser.add_option('--all', '-a', action='store_true', | 403 parser.add_option('--all', '-a', action='store_true', |
| 338 help='Unlock all repository caches') | 404 help='Unlock all repository caches') |
| 339 options, args = parser.parse_args(args) | 405 options, args = parser.parse_args(args) |
| 340 if len(args) > 1 or (len(args) == 0 and not options.all): | 406 if len(args) > 1 or (len(args) == 0 and not options.all): |
| 341 parser.error('git cache unlock takes exactly one repo url, or --all') | 407 parser.error('git cache unlock takes exactly one repo url, or --all') |
| 342 | 408 |
| 409 repo_dirs = [] | |
| 343 if not options.all: | 410 if not options.all: |
| 344 url = args[0] | 411 url = args[0] |
| 345 repo_dirs = [os.path.join(options.cache_dir, UrlToCacheDir(url))] | 412 repo_dirs.append(Mirror(url).mirror_path) |
| 346 else: | 413 else: |
| 347 repo_dirs = [os.path.join(options.cache_dir, path) | 414 cachepath = Mirror.GetCachePath() |
| 348 for path in os.listdir(options.cache_dir) | 415 repo_dirs = [os.path.join(cachepath, path) |
| 349 if os.path.isdir(os.path.join(options.cache_dir, path))] | 416 for path in os.listdir(cachepath) |
| 350 repo_dirs.extend([os.path.join(options.cache_dir, | 417 if os.path.isdir(os.path.join(cachepath, path))] |
| 418 repo_dirs.extend([os.path.join(cachepath, | |
| 351 lockfile.replace('.lock', '')) | 419 lockfile.replace('.lock', '')) |
| 352 for lockfile in os.listdir(options.cache_dir) | 420 for lockfile in os.listdir(cachepath) |
| 353 if os.path.isfile(os.path.join(options.cache_dir, | 421 if os.path.isfile(os.path.join(cachepath, |
| 354 lockfile)) | 422 lockfile)) |
| 355 and lockfile.endswith('.lock') | 423 and lockfile.endswith('.lock') |
| 356 and os.path.join(options.cache_dir, lockfile) | 424 and os.path.join(cachepath, lockfile) |
| 357 not in repo_dirs]) | 425 not in repo_dirs]) |
| 358 lockfiles = [repo_dir + '.lock' for repo_dir in repo_dirs | 426 lockfiles = [repo_dir + '.lock' for repo_dir in repo_dirs |
| 359 if os.path.exists(repo_dir + '.lock')] | 427 if os.path.exists(repo_dir + '.lock')] |
| 360 | 428 |
| 361 if not options.force: | 429 if not options.force: |
| 362 parser.error('git cache unlock requires -f|--force to do anything. ' | 430 parser.error('git cache unlock requires -f|--force to do anything. ' |
| 363 'Refusing to unlock the following repo caches: ' | 431 'Refusing to unlock the following repo caches: ' |
| 364 ', '.join(lockfiles)) | 432 ', '.join(lockfiles)) |
| 365 | 433 |
| 366 unlocked = [] | 434 unlocked_repos = [] |
| 367 untouched = [] | 435 untouched_repos = [] |
| 368 for repo_dir in repo_dirs: | 436 for repo_dir in repo_dirs: |
| 369 lf = Lockfile(repo_dir) | 437 lf = Lockfile(repo_dir) |
| 370 config_lock = os.path.join(repo_dir, 'config.lock') | 438 config_lock = os.path.join(repo_dir, 'config.lock') |
| 371 unlocked = False | 439 unlocked = False |
| 372 if os.path.exists(config_lock): | 440 if os.path.exists(config_lock): |
| 373 os.remove(config_lock) | 441 os.remove(config_lock) |
| 374 unlocked = True | 442 unlocked = True |
| 375 if lf.break_lock(): | 443 if lf.break_lock(): |
| 376 unlocked = True | 444 unlocked = True |
| 377 | 445 |
| 378 if unlocked: | 446 if unlocked: |
| 379 unlocked.append(repo_dir) | 447 unlocked_repos.append(repo_dir) |
| 380 else: | 448 else: |
| 381 untouched.append(repo_dir) | 449 untouched_repos.append(repo_dir) |
| 382 | 450 |
| 383 if unlocked: | 451 if unlocked_repos: |
| 384 logging.info('Broke locks on these caches: %s' % unlocked) | 452 logging.info('Broke locks on these caches:\n %s' % '\n '.join( |
| 385 if untouched: | 453 unlocked_repos)) |
| 386 logging.debug('Did not touch these caches: %s' % untouched) | 454 if untouched_repos: |
| 455 logging.debug('Did not touch these caches:\n %s' % '\n '.join( | |
| 456 untouched_repos)) | |
| 387 | 457 |
| 388 | 458 |
| 389 class OptionParser(optparse.OptionParser): | 459 class OptionParser(optparse.OptionParser): |
| 390 """Wrapper class for OptionParser to handle global options.""" | 460 """Wrapper class for OptionParser to handle global options.""" |
| 391 | 461 |
| 392 def __init__(self, *args, **kwargs): | 462 def __init__(self, *args, **kwargs): |
| 393 optparse.OptionParser.__init__(self, *args, prog='git cache', **kwargs) | 463 optparse.OptionParser.__init__(self, *args, prog='git cache', **kwargs) |
| 394 self.add_option('-c', '--cache-dir', | 464 self.add_option('-c', '--cache-dir', |
| 395 help='Path to the directory containing the cache') | 465 help='Path to the directory containing the cache') |
| 396 self.add_option('-v', '--verbose', action='count', default=0, | 466 self.add_option('-v', '--verbose', action='count', default=0, |
| 397 help='Increase verbosity (can be passed multiple times)') | 467 help='Increase verbosity (can be passed multiple times)') |
| 398 | 468 |
| 399 def parse_args(self, args=None, values=None): | 469 def parse_args(self, args=None, values=None): |
| 400 options, args = optparse.OptionParser.parse_args(self, args, values) | 470 options, args = optparse.OptionParser.parse_args(self, args, values) |
| 401 | 471 |
| 402 try: | 472 try: |
| 403 global_cache_dir = subprocess.check_output( | 473 global_cache_dir = Mirror.GetCachePath() |
| 404 [GIT_EXECUTABLE, 'config', '--global', 'cache.cachepath']).strip() | 474 except RuntimeError: |
| 405 if options.cache_dir: | 475 global_cache_dir = None |
| 406 if global_cache_dir and ( | 476 if options.cache_dir: |
| 407 os.path.abspath(options.cache_dir) != | 477 if global_cache_dir and ( |
| 408 os.path.abspath(global_cache_dir)): | 478 os.path.abspath(options.cache_dir) != |
| 409 logging.warn('Overriding globally-configured cache directory.') | 479 os.path.abspath(global_cache_dir)): |
| 410 else: | 480 logging.warn('Overriding globally-configured cache directory.') |
| 411 options.cache_dir = global_cache_dir | 481 Mirror.SetCachePath(options.cache_dir) |
| 412 except subprocess.CalledProcessError: | |
| 413 if not options.cache_dir: | |
| 414 self.error('No cache directory specified on command line ' | |
| 415 'or in cache.cachepath.') | |
| 416 options.cache_dir = os.path.abspath(options.cache_dir) | |
| 417 | 482 |
| 418 levels = [logging.WARNING, logging.INFO, logging.DEBUG] | 483 levels = [logging.WARNING, logging.INFO, logging.DEBUG] |
| 419 logging.basicConfig(level=levels[min(options.verbose, len(levels) - 1)]) | 484 logging.basicConfig(level=levels[min(options.verbose, len(levels) - 1)]) |
| 420 | 485 |
| 421 return options, args | 486 return options, args |
| 422 | 487 |
| 423 | 488 |
| 424 def main(argv): | 489 def main(argv): |
| 425 dispatcher = subcommand.CommandDispatcher(__name__) | 490 dispatcher = subcommand.CommandDispatcher(__name__) |
| 426 return dispatcher.execute(OptionParser(), argv) | 491 return dispatcher.execute(OptionParser(), argv) |
| 427 | 492 |
| 428 | 493 |
| 429 if __name__ == '__main__': | 494 if __name__ == '__main__': |
| 430 sys.exit(main(sys.argv[1:])) | 495 sys.exit(main(sys.argv[1:])) |
| OLD | NEW |