OLD | NEW |
| (Empty) |
1 #!/usr/bin/env python | |
2 # Copyright (c) 2012 The Chromium Authors. All rights reserved. | |
3 # Use of this source code is governed by a BSD-style license that can be | |
4 # found in the LICENSE file. | |
5 | |
6 """Front end tool to operate on .isolate files. | |
7 | |
8 This includes creating, merging or compiling them to generate a .isolated file. | |
9 | |
10 See more information at | |
11 https://code.google.com/p/swarming/wiki/IsolateDesign | |
12 https://code.google.com/p/swarming/wiki/IsolateUserGuide | |
13 """ | |
14 # Run ./isolate.py --help for more detailed information. | |
15 | |
16 import ast | |
17 import copy | |
18 import itertools | |
19 import logging | |
20 import optparse | |
21 import os | |
22 import posixpath | |
23 import re | |
24 import stat | |
25 import subprocess | |
26 import sys | |
27 | |
28 import isolateserver | |
29 import run_isolated | |
30 import trace_inputs | |
31 | |
32 # Import here directly so isolate is easier to use as a library. | |
33 from run_isolated import get_flavor | |
34 | |
35 from third_party import colorama | |
36 from third_party.depot_tools import fix_encoding | |
37 from third_party.depot_tools import subcommand | |
38 | |
39 from utils import file_path | |
40 from utils import tools | |
41 from utils import short_expression_finder | |
42 | |
43 | |
44 __version__ = '0.1.1' | |
45 | |
46 | |
47 PATH_VARIABLES = ('DEPTH', 'PRODUCT_DIR') | |
48 | |
49 # Files that should be 0-length when mapped. | |
50 KEY_TOUCHED = 'isolate_dependency_touched' | |
51 # Files that should be tracked by the build tool. | |
52 KEY_TRACKED = 'isolate_dependency_tracked' | |
53 # Files that should not be tracked by the build tool. | |
54 KEY_UNTRACKED = 'isolate_dependency_untracked' | |
55 | |
56 | |
57 class ExecutionError(Exception): | |
58 """A generic error occurred.""" | |
59 def __str__(self): | |
60 return self.args[0] | |
61 | |
62 | |
63 ### Path handling code. | |
64 | |
65 | |
66 DEFAULT_BLACKLIST = ( | |
67 # Temporary vim or python files. | |
68 r'^.+\.(?:pyc|swp)$', | |
69 # .git or .svn directory. | |
70 r'^(?:.+' + re.escape(os.path.sep) + r'|)\.(?:git|svn)$', | |
71 ) | |
72 | |
73 | |
74 # Chromium-specific. | |
75 DEFAULT_BLACKLIST += ( | |
76 r'^.+\.(?:run_test_cases)$', | |
77 r'^(?:.+' + re.escape(os.path.sep) + r'|)testserver\.log$', | |
78 ) | |
79 | |
80 | |
81 def relpath(path, root): | |
82 """os.path.relpath() that keeps trailing os.path.sep.""" | |
83 out = os.path.relpath(path, root) | |
84 if path.endswith(os.path.sep): | |
85 out += os.path.sep | |
86 return out | |
87 | |
88 | |
89 def safe_relpath(filepath, basepath): | |
90 """Do not throw on Windows when filepath and basepath are on different drives. | |
91 | |
92 Different than relpath() above since this one doesn't keep the trailing | |
93 os.path.sep and it swallows exceptions on Windows and return the original | |
94 absolute path in the case of different drives. | |
95 """ | |
96 try: | |
97 return os.path.relpath(filepath, basepath) | |
98 except ValueError: | |
99 assert sys.platform == 'win32' | |
100 return filepath | |
101 | |
102 | |
103 def normpath(path): | |
104 """os.path.normpath() that keeps trailing os.path.sep.""" | |
105 out = os.path.normpath(path) | |
106 if path.endswith(os.path.sep): | |
107 out += os.path.sep | |
108 return out | |
109 | |
110 | |
111 def posix_relpath(path, root): | |
112 """posix.relpath() that keeps trailing slash.""" | |
113 out = posixpath.relpath(path, root) | |
114 if path.endswith('/'): | |
115 out += '/' | |
116 return out | |
117 | |
118 | |
119 def cleanup_path(x): | |
120 """Cleans up a relative path. Converts any os.path.sep to '/' on Windows.""" | |
121 if x: | |
122 x = x.rstrip(os.path.sep).replace(os.path.sep, '/') | |
123 if x == '.': | |
124 x = '' | |
125 if x: | |
126 x += '/' | |
127 return x | |
128 | |
129 | |
130 def is_url(path): | |
131 return bool(re.match(r'^https?://.+$', path)) | |
132 | |
133 | |
134 def path_starts_with(prefix, path): | |
135 """Returns true if the components of the path |prefix| are the same as the | |
136 initial components of |path| (or all of the components of |path|). The paths | |
137 must be absolute. | |
138 """ | |
139 assert os.path.isabs(prefix) and os.path.isabs(path) | |
140 prefix = os.path.normpath(prefix) | |
141 path = os.path.normpath(path) | |
142 assert prefix == file_path.get_native_path_case(prefix), prefix | |
143 assert path == file_path.get_native_path_case(path), path | |
144 prefix = prefix.rstrip(os.path.sep) + os.path.sep | |
145 path = path.rstrip(os.path.sep) + os.path.sep | |
146 return path.startswith(prefix) | |
147 | |
148 | |
149 def fix_native_path_case(root, path): | |
150 """Ensures that each component of |path| has the proper native case by | |
151 iterating slowly over the directory elements of |path|.""" | |
152 native_case_path = root | |
153 for raw_part in path.split(os.sep): | |
154 if not raw_part or raw_part == '.': | |
155 break | |
156 | |
157 part = file_path.find_item_native_case(native_case_path, raw_part) | |
158 if not part: | |
159 raise isolateserver.MappingError( | |
160 'Input file %s doesn\'t exist' % | |
161 os.path.join(native_case_path, raw_part)) | |
162 native_case_path = os.path.join(native_case_path, part) | |
163 | |
164 return os.path.normpath(native_case_path) | |
165 | |
166 | |
167 def expand_symlinks(indir, relfile): | |
168 """Follows symlinks in |relfile|, but treating symlinks that point outside the | |
169 build tree as if they were ordinary directories/files. Returns the final | |
170 symlink-free target and a list of paths to symlinks encountered in the | |
171 process. | |
172 | |
173 The rule about symlinks outside the build tree is for the benefit of the | |
174 Chromium OS ebuild, which symlinks the output directory to an unrelated path | |
175 in the chroot. | |
176 | |
177 Fails when a directory loop is detected, although in theory we could support | |
178 that case. | |
179 """ | |
180 is_directory = relfile.endswith(os.path.sep) | |
181 done = indir | |
182 todo = relfile.strip(os.path.sep) | |
183 symlinks = [] | |
184 | |
185 while todo: | |
186 pre_symlink, symlink, post_symlink = file_path.split_at_symlink( | |
187 done, todo) | |
188 if not symlink: | |
189 todo = fix_native_path_case(done, todo) | |
190 done = os.path.join(done, todo) | |
191 break | |
192 symlink_path = os.path.join(done, pre_symlink, symlink) | |
193 post_symlink = post_symlink.lstrip(os.path.sep) | |
194 # readlink doesn't exist on Windows. | |
195 # pylint: disable=E1101 | |
196 target = os.path.normpath(os.path.join(done, pre_symlink)) | |
197 symlink_target = os.readlink(symlink_path) | |
198 if os.path.isabs(symlink_target): | |
199 # Absolute path are considered a normal directories. The use case is | |
200 # generally someone who puts the output directory on a separate drive. | |
201 target = symlink_target | |
202 else: | |
203 # The symlink itself could be using the wrong path case. | |
204 target = fix_native_path_case(target, symlink_target) | |
205 | |
206 if not os.path.exists(target): | |
207 raise isolateserver.MappingError( | |
208 'Symlink target doesn\'t exist: %s -> %s' % (symlink_path, target)) | |
209 target = file_path.get_native_path_case(target) | |
210 if not path_starts_with(indir, target): | |
211 done = symlink_path | |
212 todo = post_symlink | |
213 continue | |
214 if path_starts_with(target, symlink_path): | |
215 raise isolateserver.MappingError( | |
216 'Can\'t map recursive symlink reference %s -> %s' % | |
217 (symlink_path, target)) | |
218 logging.info('Found symlink: %s -> %s', symlink_path, target) | |
219 symlinks.append(os.path.relpath(symlink_path, indir)) | |
220 # Treat the common prefix of the old and new paths as done, and start | |
221 # scanning again. | |
222 target = target.split(os.path.sep) | |
223 symlink_path = symlink_path.split(os.path.sep) | |
224 prefix_length = 0 | |
225 for target_piece, symlink_path_piece in zip(target, symlink_path): | |
226 if target_piece == symlink_path_piece: | |
227 prefix_length += 1 | |
228 else: | |
229 break | |
230 done = os.path.sep.join(target[:prefix_length]) | |
231 todo = os.path.join( | |
232 os.path.sep.join(target[prefix_length:]), post_symlink) | |
233 | |
234 relfile = os.path.relpath(done, indir) | |
235 relfile = relfile.rstrip(os.path.sep) + is_directory * os.path.sep | |
236 return relfile, symlinks | |
237 | |
238 | |
239 def expand_directory_and_symlink(indir, relfile, blacklist, follow_symlinks): | |
240 """Expands a single input. It can result in multiple outputs. | |
241 | |
242 This function is recursive when relfile is a directory. | |
243 | |
244 Note: this code doesn't properly handle recursive symlink like one created | |
245 with: | |
246 ln -s .. foo | |
247 """ | |
248 if os.path.isabs(relfile): | |
249 raise isolateserver.MappingError( | |
250 'Can\'t map absolute path %s' % relfile) | |
251 | |
252 infile = normpath(os.path.join(indir, relfile)) | |
253 if not infile.startswith(indir): | |
254 raise isolateserver.MappingError( | |
255 'Can\'t map file %s outside %s' % (infile, indir)) | |
256 | |
257 filepath = os.path.join(indir, relfile) | |
258 native_filepath = file_path.get_native_path_case(filepath) | |
259 if filepath != native_filepath: | |
260 # Special case './'. | |
261 if filepath != native_filepath + '.' + os.path.sep: | |
262 # Give up enforcing strict path case on OSX. Really, it's that sad. The | |
263 # case where it happens is very specific and hard to reproduce: | |
264 # get_native_path_case( | |
265 # u'Foo.framework/Versions/A/Resources/Something.nib') will return | |
266 # u'Foo.framework/Versions/A/resources/Something.nib', e.g. lowercase 'r'. | |
267 # | |
268 # Note that this is really something deep in OSX because running | |
269 # ls Foo.framework/Versions/A | |
270 # will print out 'Resources', while file_path.get_native_path_case() | |
271 # returns a lower case 'r'. | |
272 # | |
273 # So *something* is happening under the hood resulting in the command 'ls' | |
274 # and Carbon.File.FSPathMakeRef('path').FSRefMakePath() to disagree. We | |
275 # have no idea why. | |
276 if sys.platform != 'darwin': | |
277 raise isolateserver.MappingError( | |
278 'File path doesn\'t equal native file path\n%s != %s' % | |
279 (filepath, native_filepath)) | |
280 | |
281 symlinks = [] | |
282 if follow_symlinks: | |
283 relfile, symlinks = expand_symlinks(indir, relfile) | |
284 | |
285 if relfile.endswith(os.path.sep): | |
286 if not os.path.isdir(infile): | |
287 raise isolateserver.MappingError( | |
288 '%s is not a directory but ends with "%s"' % (infile, os.path.sep)) | |
289 | |
290 # Special case './'. | |
291 if relfile.startswith('.' + os.path.sep): | |
292 relfile = relfile[2:] | |
293 outfiles = symlinks | |
294 try: | |
295 for filename in os.listdir(infile): | |
296 inner_relfile = os.path.join(relfile, filename) | |
297 if blacklist(inner_relfile): | |
298 continue | |
299 if os.path.isdir(os.path.join(indir, inner_relfile)): | |
300 inner_relfile += os.path.sep | |
301 outfiles.extend( | |
302 expand_directory_and_symlink(indir, inner_relfile, blacklist, | |
303 follow_symlinks)) | |
304 return outfiles | |
305 except OSError as e: | |
306 raise isolateserver.MappingError( | |
307 'Unable to iterate over directory %s.\n%s' % (infile, e)) | |
308 else: | |
309 # Always add individual files even if they were blacklisted. | |
310 if os.path.isdir(infile): | |
311 raise isolateserver.MappingError( | |
312 'Input directory %s must have a trailing slash' % infile) | |
313 | |
314 if not os.path.isfile(infile): | |
315 raise isolateserver.MappingError( | |
316 'Input file %s doesn\'t exist' % infile) | |
317 | |
318 return symlinks + [relfile] | |
319 | |
320 | |
321 def expand_directories_and_symlinks(indir, infiles, blacklist, | |
322 follow_symlinks, ignore_broken_items): | |
323 """Expands the directories and the symlinks, applies the blacklist and | |
324 verifies files exist. | |
325 | |
326 Files are specified in os native path separator. | |
327 """ | |
328 outfiles = [] | |
329 for relfile in infiles: | |
330 try: | |
331 outfiles.extend(expand_directory_and_symlink(indir, relfile, blacklist, | |
332 follow_symlinks)) | |
333 except isolateserver.MappingError as e: | |
334 if ignore_broken_items: | |
335 logging.info('warning: %s', e) | |
336 else: | |
337 raise | |
338 return outfiles | |
339 | |
340 | |
341 def recreate_tree(outdir, indir, infiles, action, as_hash): | |
342 """Creates a new tree with only the input files in it. | |
343 | |
344 Arguments: | |
345 outdir: Output directory to create the files in. | |
346 indir: Root directory the infiles are based in. | |
347 infiles: dict of files to map from |indir| to |outdir|. | |
348 action: One of accepted action of run_isolated.link_file(). | |
349 as_hash: Output filename is the hash instead of relfile. | |
350 """ | |
351 logging.info( | |
352 'recreate_tree(outdir=%s, indir=%s, files=%d, action=%s, as_hash=%s)' % | |
353 (outdir, indir, len(infiles), action, as_hash)) | |
354 | |
355 assert os.path.isabs(outdir) and outdir == os.path.normpath(outdir), outdir | |
356 if not os.path.isdir(outdir): | |
357 logging.info('Creating %s' % outdir) | |
358 os.makedirs(outdir) | |
359 | |
360 for relfile, metadata in infiles.iteritems(): | |
361 infile = os.path.join(indir, relfile) | |
362 if as_hash: | |
363 # Do the hashtable specific checks. | |
364 if 'l' in metadata: | |
365 # Skip links when storing a hashtable. | |
366 continue | |
367 outfile = os.path.join(outdir, metadata['h']) | |
368 if os.path.isfile(outfile): | |
369 # Just do a quick check that the file size matches. No need to stat() | |
370 # again the input file, grab the value from the dict. | |
371 if not 's' in metadata: | |
372 raise isolateserver.MappingError( | |
373 'Misconfigured item %s: %s' % (relfile, metadata)) | |
374 if metadata['s'] == os.stat(outfile).st_size: | |
375 continue | |
376 else: | |
377 logging.warn('Overwritting %s' % metadata['h']) | |
378 os.remove(outfile) | |
379 else: | |
380 outfile = os.path.join(outdir, relfile) | |
381 outsubdir = os.path.dirname(outfile) | |
382 if not os.path.isdir(outsubdir): | |
383 os.makedirs(outsubdir) | |
384 | |
385 # TODO(csharp): Fix crbug.com/150823 and enable the touched logic again. | |
386 # if metadata.get('T') == True: | |
387 # open(outfile, 'ab').close() | |
388 if 'l' in metadata: | |
389 pointed = metadata['l'] | |
390 logging.debug('Symlink: %s -> %s' % (outfile, pointed)) | |
391 # symlink doesn't exist on Windows. | |
392 os.symlink(pointed, outfile) # pylint: disable=E1101 | |
393 else: | |
394 run_isolated.link_file(outfile, infile, action) | |
395 | |
396 | |
397 def process_input(filepath, prevdict, read_only, flavor, algo): | |
398 """Processes an input file, a dependency, and return meta data about it. | |
399 | |
400 Arguments: | |
401 - filepath: File to act on. | |
402 - prevdict: the previous dictionary. It is used to retrieve the cached sha-1 | |
403 to skip recalculating the hash. | |
404 - read_only: If True, the file mode is manipulated. In practice, only save | |
405 one of 4 modes: 0755 (rwx), 0644 (rw), 0555 (rx), 0444 (r). On | |
406 windows, mode is not set since all files are 'executable' by | |
407 default. | |
408 - algo: Hashing algorithm used. | |
409 | |
410 Behaviors: | |
411 - Retrieves the file mode, file size, file timestamp, file link | |
412 destination if it is a file link and calcultate the SHA-1 of the file's | |
413 content if the path points to a file and not a symlink. | |
414 """ | |
415 out = {} | |
416 # TODO(csharp): Fix crbug.com/150823 and enable the touched logic again. | |
417 # if prevdict.get('T') == True: | |
418 # # The file's content is ignored. Skip the time and hard code mode. | |
419 # if get_flavor() != 'win': | |
420 # out['m'] = stat.S_IRUSR | stat.S_IRGRP | |
421 # out['s'] = 0 | |
422 # out['h'] = algo().hexdigest() | |
423 # out['T'] = True | |
424 # return out | |
425 | |
426 # Always check the file stat and check if it is a link. The timestamp is used | |
427 # to know if the file's content/symlink destination should be looked into. | |
428 # E.g. only reuse from prevdict if the timestamp hasn't changed. | |
429 # There is the risk of the file's timestamp being reset to its last value | |
430 # manually while its content changed. We don't protect against that use case. | |
431 try: | |
432 filestats = os.lstat(filepath) | |
433 except OSError: | |
434 # The file is not present. | |
435 raise isolateserver.MappingError('%s is missing' % filepath) | |
436 is_link = stat.S_ISLNK(filestats.st_mode) | |
437 | |
438 if flavor != 'win': | |
439 # Ignore file mode on Windows since it's not really useful there. | |
440 filemode = stat.S_IMODE(filestats.st_mode) | |
441 # Remove write access for group and all access to 'others'. | |
442 filemode &= ~(stat.S_IWGRP | stat.S_IRWXO) | |
443 if read_only: | |
444 filemode &= ~stat.S_IWUSR | |
445 if filemode & stat.S_IXUSR: | |
446 filemode |= stat.S_IXGRP | |
447 else: | |
448 filemode &= ~stat.S_IXGRP | |
449 if not is_link: | |
450 out['m'] = filemode | |
451 | |
452 # Used to skip recalculating the hash or link destination. Use the most recent | |
453 # update time. | |
454 # TODO(maruel): Save it in the .state file instead of .isolated so the | |
455 # .isolated file is deterministic. | |
456 out['t'] = int(round(filestats.st_mtime)) | |
457 | |
458 if not is_link: | |
459 out['s'] = filestats.st_size | |
460 # If the timestamp wasn't updated and the file size is still the same, carry | |
461 # on the sha-1. | |
462 if (prevdict.get('t') == out['t'] and | |
463 prevdict.get('s') == out['s']): | |
464 # Reuse the previous hash if available. | |
465 out['h'] = prevdict.get('h') | |
466 if not out.get('h'): | |
467 out['h'] = isolateserver.hash_file(filepath, algo) | |
468 else: | |
469 # If the timestamp wasn't updated, carry on the link destination. | |
470 if prevdict.get('t') == out['t']: | |
471 # Reuse the previous link destination if available. | |
472 out['l'] = prevdict.get('l') | |
473 if out.get('l') is None: | |
474 # The link could be in an incorrect path case. In practice, this only | |
475 # happen on OSX on case insensitive HFS. | |
476 # TODO(maruel): It'd be better if it was only done once, in | |
477 # expand_directory_and_symlink(), so it would not be necessary to do again | |
478 # here. | |
479 symlink_value = os.readlink(filepath) # pylint: disable=E1101 | |
480 filedir = file_path.get_native_path_case(os.path.dirname(filepath)) | |
481 native_dest = fix_native_path_case(filedir, symlink_value) | |
482 out['l'] = os.path.relpath(native_dest, filedir) | |
483 return out | |
484 | |
485 | |
486 ### Variable stuff. | |
487 | |
488 | |
489 def isolatedfile_to_state(filename): | |
490 """Replaces the file's extension.""" | |
491 return filename + '.state' | |
492 | |
493 | |
494 def determine_root_dir(relative_root, infiles): | |
495 """For a list of infiles, determines the deepest root directory that is | |
496 referenced indirectly. | |
497 | |
498 All arguments must be using os.path.sep. | |
499 """ | |
500 # The trick used to determine the root directory is to look at "how far" back | |
501 # up it is looking up. | |
502 deepest_root = relative_root | |
503 for i in infiles: | |
504 x = relative_root | |
505 while i.startswith('..' + os.path.sep): | |
506 i = i[3:] | |
507 assert not i.startswith(os.path.sep) | |
508 x = os.path.dirname(x) | |
509 if deepest_root.startswith(x): | |
510 deepest_root = x | |
511 logging.debug( | |
512 'determine_root_dir(%s, %d files) -> %s' % ( | |
513 relative_root, len(infiles), deepest_root)) | |
514 return deepest_root | |
515 | |
516 | |
517 def replace_variable(part, variables): | |
518 m = re.match(r'<\(([A-Z_]+)\)', part) | |
519 if m: | |
520 if m.group(1) not in variables: | |
521 raise ExecutionError( | |
522 'Variable "%s" was not found in %s.\nDid you forget to specify ' | |
523 '--variable?' % (m.group(1), variables)) | |
524 return variables[m.group(1)] | |
525 return part | |
526 | |
527 | |
528 def process_variables(cwd, variables, relative_base_dir): | |
529 """Processes path variables as a special case and returns a copy of the dict. | |
530 | |
531 For each 'path' variable: first normalizes it based on |cwd|, verifies it | |
532 exists then sets it as relative to relative_base_dir. | |
533 """ | |
534 relative_base_dir = file_path.get_native_path_case(relative_base_dir) | |
535 variables = variables.copy() | |
536 for i in PATH_VARIABLES: | |
537 if i not in variables: | |
538 continue | |
539 variable = variables[i].strip() | |
540 # Variables could contain / or \ on windows. Always normalize to | |
541 # os.path.sep. | |
542 variable = variable.replace('/', os.path.sep) | |
543 variable = os.path.join(cwd, variable) | |
544 variable = os.path.normpath(variable) | |
545 variable = file_path.get_native_path_case(variable) | |
546 if not os.path.isdir(variable): | |
547 raise ExecutionError('%s=%s is not a directory' % (i, variable)) | |
548 | |
549 # All variables are relative to the .isolate file. | |
550 variable = os.path.relpath(variable, relative_base_dir) | |
551 logging.debug( | |
552 'Translated variable %s from %s to %s', i, variables[i], variable) | |
553 variables[i] = variable | |
554 return variables | |
555 | |
556 | |
557 def eval_variables(item, variables): | |
558 """Replaces the .isolate variables in a string item. | |
559 | |
560 Note that the .isolate format is a subset of the .gyp dialect. | |
561 """ | |
562 return ''.join( | |
563 replace_variable(p, variables) for p in re.split(r'(<\([A-Z_]+\))', item)) | |
564 | |
565 | |
566 def classify_files(root_dir, tracked, untracked): | |
567 """Converts the list of files into a .isolate 'variables' dictionary. | |
568 | |
569 Arguments: | |
570 - tracked: list of files names to generate a dictionary out of that should | |
571 probably be tracked. | |
572 - untracked: list of files names that must not be tracked. | |
573 """ | |
574 # These directories are not guaranteed to be always present on every builder. | |
575 OPTIONAL_DIRECTORIES = ( | |
576 'test/data/plugin', | |
577 'third_party/WebKit/LayoutTests', | |
578 ) | |
579 | |
580 new_tracked = [] | |
581 new_untracked = list(untracked) | |
582 | |
583 def should_be_tracked(filepath): | |
584 """Returns True if it is a file without whitespace in a non-optional | |
585 directory that has no symlink in its path. | |
586 """ | |
587 if filepath.endswith('/'): | |
588 return False | |
589 if ' ' in filepath: | |
590 return False | |
591 if any(i in filepath for i in OPTIONAL_DIRECTORIES): | |
592 return False | |
593 # Look if any element in the path is a symlink. | |
594 split = filepath.split('/') | |
595 for i in range(len(split)): | |
596 if os.path.islink(os.path.join(root_dir, '/'.join(split[:i+1]))): | |
597 return False | |
598 return True | |
599 | |
600 for filepath in sorted(tracked): | |
601 if should_be_tracked(filepath): | |
602 new_tracked.append(filepath) | |
603 else: | |
604 # Anything else. | |
605 new_untracked.append(filepath) | |
606 | |
607 variables = {} | |
608 if new_tracked: | |
609 variables[KEY_TRACKED] = sorted(new_tracked) | |
610 if new_untracked: | |
611 variables[KEY_UNTRACKED] = sorted(new_untracked) | |
612 return variables | |
613 | |
614 | |
615 def chromium_fix(f, variables): | |
616 """Fixes an isolate dependnecy with Chromium-specific fixes.""" | |
617 # Skip log in PRODUCT_DIR. Note that these are applied on '/' style path | |
618 # separator. | |
619 LOG_FILE = re.compile(r'^\<\(PRODUCT_DIR\)\/[^\/]+\.log$') | |
620 # Ignored items. | |
621 IGNORED_ITEMS = ( | |
622 # http://crbug.com/160539, on Windows, it's in chrome/. | |
623 'Media Cache/', | |
624 'chrome/Media Cache/', | |
625 # 'First Run' is not created by the compile, but by the test itself. | |
626 '<(PRODUCT_DIR)/First Run') | |
627 | |
628 # Blacklist logs and other unimportant files. | |
629 if LOG_FILE.match(f) or f in IGNORED_ITEMS: | |
630 logging.debug('Ignoring %s', f) | |
631 return None | |
632 | |
633 EXECUTABLE = re.compile( | |
634 r'^(\<\(PRODUCT_DIR\)\/[^\/\.]+)' + | |
635 re.escape(variables.get('EXECUTABLE_SUFFIX', '')) + | |
636 r'$') | |
637 match = EXECUTABLE.match(f) | |
638 if match: | |
639 return match.group(1) + '<(EXECUTABLE_SUFFIX)' | |
640 | |
641 if sys.platform == 'darwin': | |
642 # On OSX, the name of the output is dependent on gyp define, it can be | |
643 # 'Google Chrome.app' or 'Chromium.app', same for 'XXX | |
644 # Framework.framework'. Furthermore, they are versioned with a gyp | |
645 # variable. To lower the complexity of the .isolate file, remove all the | |
646 # individual entries that show up under any of the 4 entries and replace | |
647 # them with the directory itself. Overall, this results in a bit more | |
648 # files than strictly necessary. | |
649 OSX_BUNDLES = ( | |
650 '<(PRODUCT_DIR)/Chromium Framework.framework/', | |
651 '<(PRODUCT_DIR)/Chromium.app/', | |
652 '<(PRODUCT_DIR)/Google Chrome Framework.framework/', | |
653 '<(PRODUCT_DIR)/Google Chrome.app/', | |
654 ) | |
655 for prefix in OSX_BUNDLES: | |
656 if f.startswith(prefix): | |
657 # Note this result in duplicate values, so the a set() must be used to | |
658 # remove duplicates. | |
659 return prefix | |
660 return f | |
661 | |
662 | |
663 def generate_simplified( | |
664 tracked, untracked, touched, root_dir, variables, relative_cwd, | |
665 trace_blacklist): | |
666 """Generates a clean and complete .isolate 'variables' dictionary. | |
667 | |
668 Cleans up and extracts only files from within root_dir then processes | |
669 variables and relative_cwd. | |
670 """ | |
671 root_dir = os.path.realpath(root_dir) | |
672 logging.info( | |
673 'generate_simplified(%d files, %s, %s, %s)' % | |
674 (len(tracked) + len(untracked) + len(touched), | |
675 root_dir, variables, relative_cwd)) | |
676 | |
677 # Preparation work. | |
678 relative_cwd = cleanup_path(relative_cwd) | |
679 assert not os.path.isabs(relative_cwd), relative_cwd | |
680 # Creates the right set of variables here. We only care about PATH_VARIABLES. | |
681 path_variables = dict( | |
682 ('<(%s)' % k, variables[k].replace(os.path.sep, '/')) | |
683 for k in PATH_VARIABLES if k in variables) | |
684 variables = variables.copy() | |
685 variables.update(path_variables) | |
686 | |
687 # Actual work: Process the files. | |
688 # TODO(maruel): if all the files in a directory are in part tracked and in | |
689 # part untracked, the directory will not be extracted. Tracked files should be | |
690 # 'promoted' to be untracked as needed. | |
691 tracked = trace_inputs.extract_directories( | |
692 root_dir, tracked, trace_blacklist) | |
693 untracked = trace_inputs.extract_directories( | |
694 root_dir, untracked, trace_blacklist) | |
695 # touched is not compressed, otherwise it would result in files to be archived | |
696 # that we don't need. | |
697 | |
698 root_dir_posix = root_dir.replace(os.path.sep, '/') | |
699 def fix(f): | |
700 """Bases the file on the most restrictive variable.""" | |
701 # Important, GYP stores the files with / and not \. | |
702 f = f.replace(os.path.sep, '/') | |
703 logging.debug('fix(%s)' % f) | |
704 # If it's not already a variable. | |
705 if not f.startswith('<'): | |
706 # relative_cwd is usually the directory containing the gyp file. It may be | |
707 # empty if the whole directory containing the gyp file is needed. | |
708 # Use absolute paths in case cwd_dir is outside of root_dir. | |
709 # Convert the whole thing to / since it's isolate's speak. | |
710 f = posix_relpath( | |
711 posixpath.join(root_dir_posix, f), | |
712 posixpath.join(root_dir_posix, relative_cwd)) or './' | |
713 | |
714 for variable, root_path in path_variables.iteritems(): | |
715 if f.startswith(root_path): | |
716 f = variable + f[len(root_path):] | |
717 logging.debug('Converted to %s' % f) | |
718 break | |
719 return f | |
720 | |
721 def fix_all(items): | |
722 """Reduces the items to convert variables, removes unneeded items, apply | |
723 chromium-specific fixes and only return unique items. | |
724 """ | |
725 variables_converted = (fix(f.path) for f in items) | |
726 chromium_fixed = (chromium_fix(f, variables) for f in variables_converted) | |
727 return set(f for f in chromium_fixed if f) | |
728 | |
729 tracked = fix_all(tracked) | |
730 untracked = fix_all(untracked) | |
731 touched = fix_all(touched) | |
732 out = classify_files(root_dir, tracked, untracked) | |
733 if touched: | |
734 out[KEY_TOUCHED] = sorted(touched) | |
735 return out | |
736 | |
737 | |
738 def chromium_filter_flags(variables): | |
739 """Filters out build flags used in Chromium that we don't want to treat as | |
740 configuration variables. | |
741 """ | |
742 # TODO(benrg): Need a better way to determine this. | |
743 blacklist = set(PATH_VARIABLES + ('EXECUTABLE_SUFFIX', 'FLAG')) | |
744 return dict((k, v) for k, v in variables.iteritems() if k not in blacklist) | |
745 | |
746 | |
747 def generate_isolate( | |
748 tracked, untracked, touched, root_dir, variables, relative_cwd, | |
749 trace_blacklist): | |
750 """Generates a clean and complete .isolate file.""" | |
751 dependencies = generate_simplified( | |
752 tracked, untracked, touched, root_dir, variables, relative_cwd, | |
753 trace_blacklist) | |
754 config_variables = chromium_filter_flags(variables) | |
755 config_variable_names, config_values = zip( | |
756 *sorted(config_variables.iteritems())) | |
757 out = Configs(None) | |
758 # The new dependencies apply to just one configuration, namely config_values. | |
759 out.merge_dependencies(dependencies, config_variable_names, [config_values]) | |
760 return out.make_isolate_file() | |
761 | |
762 | |
763 def split_touched(files): | |
764 """Splits files that are touched vs files that are read.""" | |
765 tracked = [] | |
766 touched = [] | |
767 for f in files: | |
768 if f.size: | |
769 tracked.append(f) | |
770 else: | |
771 touched.append(f) | |
772 return tracked, touched | |
773 | |
774 | |
775 def pretty_print(variables, stdout): | |
776 """Outputs a gyp compatible list from the decoded variables. | |
777 | |
778 Similar to pprint.print() but with NIH syndrome. | |
779 """ | |
780 # Order the dictionary keys by these keys in priority. | |
781 ORDER = ( | |
782 'variables', 'condition', 'command', 'relative_cwd', 'read_only', | |
783 KEY_TRACKED, KEY_UNTRACKED) | |
784 | |
785 def sorting_key(x): | |
786 """Gives priority to 'most important' keys before the others.""" | |
787 if x in ORDER: | |
788 return str(ORDER.index(x)) | |
789 return x | |
790 | |
791 def loop_list(indent, items): | |
792 for item in items: | |
793 if isinstance(item, basestring): | |
794 stdout.write('%s\'%s\',\n' % (indent, item)) | |
795 elif isinstance(item, dict): | |
796 stdout.write('%s{\n' % indent) | |
797 loop_dict(indent + ' ', item) | |
798 stdout.write('%s},\n' % indent) | |
799 elif isinstance(item, list): | |
800 # A list inside a list will write the first item embedded. | |
801 stdout.write('%s[' % indent) | |
802 for index, i in enumerate(item): | |
803 if isinstance(i, basestring): | |
804 stdout.write( | |
805 '\'%s\', ' % i.replace('\\', '\\\\').replace('\'', '\\\'')) | |
806 elif isinstance(i, dict): | |
807 stdout.write('{\n') | |
808 loop_dict(indent + ' ', i) | |
809 if index != len(item) - 1: | |
810 x = ', ' | |
811 else: | |
812 x = '' | |
813 stdout.write('%s}%s' % (indent, x)) | |
814 else: | |
815 assert False | |
816 stdout.write('],\n') | |
817 else: | |
818 assert False | |
819 | |
820 def loop_dict(indent, items): | |
821 for key in sorted(items, key=sorting_key): | |
822 item = items[key] | |
823 stdout.write("%s'%s': " % (indent, key)) | |
824 if isinstance(item, dict): | |
825 stdout.write('{\n') | |
826 loop_dict(indent + ' ', item) | |
827 stdout.write(indent + '},\n') | |
828 elif isinstance(item, list): | |
829 stdout.write('[\n') | |
830 loop_list(indent + ' ', item) | |
831 stdout.write(indent + '],\n') | |
832 elif isinstance(item, basestring): | |
833 stdout.write( | |
834 '\'%s\',\n' % item.replace('\\', '\\\\').replace('\'', '\\\'')) | |
835 elif item in (True, False, None): | |
836 stdout.write('%s\n' % item) | |
837 else: | |
838 assert False, item | |
839 | |
840 stdout.write('{\n') | |
841 loop_dict(' ', variables) | |
842 stdout.write('}\n') | |
843 | |
844 | |
845 def union(lhs, rhs): | |
846 """Merges two compatible datastructures composed of dict/list/set.""" | |
847 assert lhs is not None or rhs is not None | |
848 if lhs is None: | |
849 return copy.deepcopy(rhs) | |
850 if rhs is None: | |
851 return copy.deepcopy(lhs) | |
852 assert type(lhs) == type(rhs), (lhs, rhs) | |
853 if hasattr(lhs, 'union'): | |
854 # Includes set, ConfigSettings and Configs. | |
855 return lhs.union(rhs) | |
856 if isinstance(lhs, dict): | |
857 return dict((k, union(lhs.get(k), rhs.get(k))) for k in set(lhs).union(rhs)) | |
858 elif isinstance(lhs, list): | |
859 # Do not go inside the list. | |
860 return lhs + rhs | |
861 assert False, type(lhs) | |
862 | |
863 | |
864 def extract_comment(content): | |
865 """Extracts file level comment.""" | |
866 out = [] | |
867 for line in content.splitlines(True): | |
868 if line.startswith('#'): | |
869 out.append(line) | |
870 else: | |
871 break | |
872 return ''.join(out) | |
873 | |
874 | |
875 def eval_content(content): | |
876 """Evaluates a python file and return the value defined in it. | |
877 | |
878 Used in practice for .isolate files. | |
879 """ | |
880 globs = {'__builtins__': None} | |
881 locs = {} | |
882 try: | |
883 value = eval(content, globs, locs) | |
884 except TypeError as e: | |
885 e.args = list(e.args) + [content] | |
886 raise | |
887 assert locs == {}, locs | |
888 assert globs == {'__builtins__': None}, globs | |
889 return value | |
890 | |
891 | |
892 def match_configs(expr, config_variables, all_configs): | |
893 """Returns the configs from |all_configs| that match the |expr|, where | |
894 the elements of |all_configs| are tuples of values for the |config_variables|. | |
895 Example: | |
896 >>> match_configs(expr = "(foo==1 or foo==2) and bar=='b'", | |
897 config_variables = ["foo", "bar"], | |
898 all_configs = [(1, 'a'), (1, 'b'), (2, 'a'), (2, 'b')]) | |
899 [(1, 'b'), (2, 'b')] | |
900 """ | |
901 return [ | |
902 config for config in all_configs | |
903 if eval(expr, dict(zip(config_variables, config))) | |
904 ] | |
905 | |
906 | |
907 def verify_variables(variables): | |
908 """Verifies the |variables| dictionary is in the expected format.""" | |
909 VALID_VARIABLES = [ | |
910 KEY_TOUCHED, | |
911 KEY_TRACKED, | |
912 KEY_UNTRACKED, | |
913 'command', | |
914 'read_only', | |
915 ] | |
916 assert isinstance(variables, dict), variables | |
917 assert set(VALID_VARIABLES).issuperset(set(variables)), variables.keys() | |
918 for name, value in variables.iteritems(): | |
919 if name == 'read_only': | |
920 assert value in (True, False, None), value | |
921 else: | |
922 assert isinstance(value, list), value | |
923 assert all(isinstance(i, basestring) for i in value), value | |
924 | |
925 | |
926 def verify_ast(expr, variables_and_values): | |
927 """Verifies that |expr| is of the form | |
928 expr ::= expr ( "or" | "and" ) expr | |
929 | identifier "==" ( string | int ) | |
930 Also collects the variable identifiers and string/int values in the dict | |
931 |variables_and_values|, in the form {'var': set([val1, val2, ...]), ...}. | |
932 """ | |
933 assert isinstance(expr, (ast.BoolOp, ast.Compare)) | |
934 if isinstance(expr, ast.BoolOp): | |
935 assert isinstance(expr.op, (ast.And, ast.Or)) | |
936 for subexpr in expr.values: | |
937 verify_ast(subexpr, variables_and_values) | |
938 else: | |
939 assert isinstance(expr.left.ctx, ast.Load) | |
940 assert len(expr.ops) == 1 | |
941 assert isinstance(expr.ops[0], ast.Eq) | |
942 var_values = variables_and_values.setdefault(expr.left.id, set()) | |
943 rhs = expr.comparators[0] | |
944 assert isinstance(rhs, (ast.Str, ast.Num)) | |
945 var_values.add(rhs.n if isinstance(rhs, ast.Num) else rhs.s) | |
946 | |
947 | |
948 def verify_condition(condition, variables_and_values): | |
949 """Verifies the |condition| dictionary is in the expected format. | |
950 See verify_ast() for the meaning of |variables_and_values|. | |
951 """ | |
952 VALID_INSIDE_CONDITION = ['variables'] | |
953 assert isinstance(condition, list), condition | |
954 assert len(condition) == 2, condition | |
955 expr, then = condition | |
956 | |
957 test_ast = compile(expr, '<condition>', 'eval', ast.PyCF_ONLY_AST) | |
958 verify_ast(test_ast.body, variables_and_values) | |
959 | |
960 assert isinstance(then, dict), then | |
961 assert set(VALID_INSIDE_CONDITION).issuperset(set(then)), then.keys() | |
962 verify_variables(then['variables']) | |
963 | |
964 | |
965 def verify_root(value, variables_and_values): | |
966 """Verifies that |value| is the parsed form of a valid .isolate file. | |
967 See verify_ast() for the meaning of |variables_and_values|. | |
968 """ | |
969 VALID_ROOTS = ['includes', 'conditions'] | |
970 assert isinstance(value, dict), value | |
971 assert set(VALID_ROOTS).issuperset(set(value)), value.keys() | |
972 | |
973 includes = value.get('includes', []) | |
974 assert isinstance(includes, list), includes | |
975 for include in includes: | |
976 assert isinstance(include, basestring), include | |
977 | |
978 conditions = value.get('conditions', []) | |
979 assert isinstance(conditions, list), conditions | |
980 for condition in conditions: | |
981 verify_condition(condition, variables_and_values) | |
982 | |
983 | |
984 def remove_weak_dependencies(values, key, item, item_configs): | |
985 """Removes any configs from this key if the item is already under a | |
986 strong key. | |
987 """ | |
988 if key == KEY_TOUCHED: | |
989 item_configs = set(item_configs) | |
990 for stronger_key in (KEY_TRACKED, KEY_UNTRACKED): | |
991 try: | |
992 item_configs -= values[stronger_key][item] | |
993 except KeyError: | |
994 pass | |
995 | |
996 return item_configs | |
997 | |
998 | |
999 def remove_repeated_dependencies(folders, key, item, item_configs): | |
1000 """Removes any configs from this key if the item is in a folder that is | |
1001 already included.""" | |
1002 | |
1003 if key in (KEY_UNTRACKED, KEY_TRACKED, KEY_TOUCHED): | |
1004 item_configs = set(item_configs) | |
1005 for (folder, configs) in folders.iteritems(): | |
1006 if folder != item and item.startswith(folder): | |
1007 item_configs -= configs | |
1008 | |
1009 return item_configs | |
1010 | |
1011 | |
1012 def get_folders(values_dict): | |
1013 """Returns a dict of all the folders in the given value_dict.""" | |
1014 return dict( | |
1015 (item, configs) for (item, configs) in values_dict.iteritems() | |
1016 if item.endswith('/') | |
1017 ) | |
1018 | |
1019 | |
1020 def invert_map(variables): | |
1021 """Converts {config: {deptype: list(depvals)}} to | |
1022 {deptype: {depval: set(configs)}}. | |
1023 """ | |
1024 KEYS = ( | |
1025 KEY_TOUCHED, | |
1026 KEY_TRACKED, | |
1027 KEY_UNTRACKED, | |
1028 'command', | |
1029 'read_only', | |
1030 ) | |
1031 out = dict((key, {}) for key in KEYS) | |
1032 for config, values in variables.iteritems(): | |
1033 for key in KEYS: | |
1034 if key == 'command': | |
1035 items = [tuple(values[key])] if key in values else [] | |
1036 elif key == 'read_only': | |
1037 items = [values[key]] if key in values else [] | |
1038 else: | |
1039 assert key in (KEY_TOUCHED, KEY_TRACKED, KEY_UNTRACKED) | |
1040 items = values.get(key, []) | |
1041 for item in items: | |
1042 out[key].setdefault(item, set()).add(config) | |
1043 return out | |
1044 | |
1045 | |
1046 def reduce_inputs(values): | |
1047 """Reduces the output of invert_map() to the strictest minimum list. | |
1048 | |
1049 Looks at each individual file and directory, maps where they are used and | |
1050 reconstructs the inverse dictionary. | |
1051 | |
1052 Returns the minimized dictionary. | |
1053 """ | |
1054 KEYS = ( | |
1055 KEY_TOUCHED, | |
1056 KEY_TRACKED, | |
1057 KEY_UNTRACKED, | |
1058 'command', | |
1059 'read_only', | |
1060 ) | |
1061 | |
1062 # Folders can only live in KEY_UNTRACKED. | |
1063 folders = get_folders(values.get(KEY_UNTRACKED, {})) | |
1064 | |
1065 out = dict((key, {}) for key in KEYS) | |
1066 for key in KEYS: | |
1067 for item, item_configs in values.get(key, {}).iteritems(): | |
1068 item_configs = remove_weak_dependencies(values, key, item, item_configs) | |
1069 item_configs = remove_repeated_dependencies( | |
1070 folders, key, item, item_configs) | |
1071 if item_configs: | |
1072 out[key][item] = item_configs | |
1073 return out | |
1074 | |
1075 | |
1076 def convert_map_to_isolate_dict(values, config_variables): | |
1077 """Regenerates back a .isolate configuration dict from files and dirs | |
1078 mappings generated from reduce_inputs(). | |
1079 """ | |
1080 # Gather a list of configurations for set inversion later. | |
1081 all_mentioned_configs = set() | |
1082 for configs_by_item in values.itervalues(): | |
1083 for configs in configs_by_item.itervalues(): | |
1084 all_mentioned_configs.update(configs) | |
1085 | |
1086 # Invert the mapping to make it dict first. | |
1087 conditions = {} | |
1088 for key in values: | |
1089 for item, configs in values[key].iteritems(): | |
1090 then = conditions.setdefault(frozenset(configs), {}) | |
1091 variables = then.setdefault('variables', {}) | |
1092 | |
1093 if item in (True, False): | |
1094 # One-off for read_only. | |
1095 variables[key] = item | |
1096 else: | |
1097 assert item | |
1098 if isinstance(item, tuple): | |
1099 # One-off for command. | |
1100 # Do not merge lists and do not sort! | |
1101 # Note that item is a tuple. | |
1102 assert key not in variables | |
1103 variables[key] = list(item) | |
1104 else: | |
1105 # The list of items (files or dirs). Append the new item and keep | |
1106 # the list sorted. | |
1107 l = variables.setdefault(key, []) | |
1108 l.append(item) | |
1109 l.sort() | |
1110 | |
1111 if all_mentioned_configs: | |
1112 config_values = map(set, zip(*all_mentioned_configs)) | |
1113 sef = short_expression_finder.ShortExpressionFinder( | |
1114 zip(config_variables, config_values)) | |
1115 | |
1116 conditions = sorted( | |
1117 [sef.get_expr(configs), then] for configs, then in conditions.iteritems()) | |
1118 return {'conditions': conditions} | |
1119 | |
1120 | |
1121 ### Internal state files. | |
1122 | |
1123 | |
1124 class ConfigSettings(object): | |
1125 """Represents the dependency variables for a single build configuration. | |
1126 The structure is immutable. | |
1127 """ | |
1128 def __init__(self, config, values): | |
1129 self.config = config | |
1130 verify_variables(values) | |
1131 self.touched = sorted(values.get(KEY_TOUCHED, [])) | |
1132 self.tracked = sorted(values.get(KEY_TRACKED, [])) | |
1133 self.untracked = sorted(values.get(KEY_UNTRACKED, [])) | |
1134 self.command = values.get('command', [])[:] | |
1135 self.read_only = values.get('read_only') | |
1136 | |
1137 def union(self, rhs): | |
1138 assert not (self.config and rhs.config) or (self.config == rhs.config) | |
1139 assert not (self.command and rhs.command) or (self.command == rhs.command) | |
1140 var = { | |
1141 KEY_TOUCHED: sorted(self.touched + rhs.touched), | |
1142 KEY_TRACKED: sorted(self.tracked + rhs.tracked), | |
1143 KEY_UNTRACKED: sorted(self.untracked + rhs.untracked), | |
1144 'command': self.command or rhs.command, | |
1145 'read_only': rhs.read_only if self.read_only is None else self.read_only, | |
1146 } | |
1147 return ConfigSettings(self.config or rhs.config, var) | |
1148 | |
1149 def flatten(self): | |
1150 out = {} | |
1151 if self.command: | |
1152 out['command'] = self.command | |
1153 if self.touched: | |
1154 out[KEY_TOUCHED] = self.touched | |
1155 if self.tracked: | |
1156 out[KEY_TRACKED] = self.tracked | |
1157 if self.untracked: | |
1158 out[KEY_UNTRACKED] = self.untracked | |
1159 if self.read_only is not None: | |
1160 out['read_only'] = self.read_only | |
1161 return out | |
1162 | |
1163 | |
1164 class Configs(object): | |
1165 """Represents a processed .isolate file. | |
1166 | |
1167 Stores the file in a processed way, split by configuration. | |
1168 """ | |
1169 def __init__(self, file_comment): | |
1170 self.file_comment = file_comment | |
1171 # The keys of by_config are tuples of values for the configuration | |
1172 # variables. The names of the variables (which must be the same for | |
1173 # every by_config key) are kept in config_variables. Initially by_config | |
1174 # is empty and we don't know what configuration variables will be used, | |
1175 # so config_variables also starts out empty. It will be set by the first | |
1176 # call to union() or merge_dependencies(). | |
1177 self.by_config = {} | |
1178 self.config_variables = () | |
1179 | |
1180 def union(self, rhs): | |
1181 """Adds variables from rhs (a Configs) to the existing variables. | |
1182 """ | |
1183 config_variables = self.config_variables | |
1184 if not config_variables: | |
1185 config_variables = rhs.config_variables | |
1186 else: | |
1187 # We can't proceed if this isn't true since we don't know the correct | |
1188 # default values for extra variables. The variables are sorted so we | |
1189 # don't need to worry about permutations. | |
1190 if rhs.config_variables and rhs.config_variables != config_variables: | |
1191 raise ExecutionError( | |
1192 'Variables in merged .isolate files do not match: %r and %r' % ( | |
1193 config_variables, rhs.config_variables)) | |
1194 | |
1195 # Takes the first file comment, prefering lhs. | |
1196 out = Configs(self.file_comment or rhs.file_comment) | |
1197 out.config_variables = config_variables | |
1198 for config in set(self.by_config) | set(rhs.by_config): | |
1199 out.by_config[config] = union( | |
1200 self.by_config.get(config), rhs.by_config.get(config)) | |
1201 return out | |
1202 | |
1203 def merge_dependencies(self, values, config_variables, configs): | |
1204 """Adds new dependencies to this object for the given configurations. | |
1205 Arguments: | |
1206 values: A variables dict as found in a .isolate file, e.g., | |
1207 {KEY_TOUCHED: [...], 'command': ...}. | |
1208 config_variables: An ordered list of configuration variables, e.g., | |
1209 ["OS", "chromeos"]. If this object already contains any dependencies, | |
1210 the configuration variables must match. | |
1211 configs: a list of tuples of values of the configuration variables, | |
1212 e.g., [("mac", 0), ("linux", 1)]. The dependencies in |values| | |
1213 are added to all of these configurations, and other configurations | |
1214 are unchanged. | |
1215 """ | |
1216 if not values: | |
1217 return | |
1218 | |
1219 if not self.config_variables: | |
1220 self.config_variables = config_variables | |
1221 else: | |
1222 # See comment in Configs.union(). | |
1223 assert self.config_variables == config_variables | |
1224 | |
1225 for config in configs: | |
1226 self.by_config[config] = union( | |
1227 self.by_config.get(config), ConfigSettings(config, values)) | |
1228 | |
1229 def flatten(self): | |
1230 """Returns a flat dictionary representation of the configuration. | |
1231 """ | |
1232 return dict((k, v.flatten()) for k, v in self.by_config.iteritems()) | |
1233 | |
1234 def make_isolate_file(self): | |
1235 """Returns a dictionary suitable for writing to a .isolate file. | |
1236 """ | |
1237 dependencies_by_config = self.flatten() | |
1238 configs_by_dependency = reduce_inputs(invert_map(dependencies_by_config)) | |
1239 return convert_map_to_isolate_dict(configs_by_dependency, | |
1240 self.config_variables) | |
1241 | |
1242 | |
1243 # TODO(benrg): Remove this function when no old-format files are left. | |
1244 def convert_old_to_new_format(value): | |
1245 """Converts from the old .isolate format, which only has one variable (OS), | |
1246 always includes 'linux', 'mac' and 'win' in the set of valid values for OS, | |
1247 and allows conditions that depend on the set of all OSes, to the new format, | |
1248 which allows any set of variables, has no hardcoded values, and only allows | |
1249 explicit positive tests of variable values. | |
1250 """ | |
1251 conditions = value.get('conditions', []) | |
1252 if 'variables' not in value and all(len(cond) == 2 for cond in conditions): | |
1253 return value # Nothing to change | |
1254 | |
1255 def parse_condition(cond): | |
1256 return re.match(r'OS=="(\w+)"\Z', cond[0]).group(1) | |
1257 | |
1258 oses = set(map(parse_condition, conditions)) | |
1259 default_oses = set(['linux', 'mac', 'win']) | |
1260 oses = sorted(oses | default_oses) | |
1261 | |
1262 def if_not_os(not_os, then): | |
1263 expr = ' or '.join('OS=="%s"' % os for os in oses if os != not_os) | |
1264 return [expr, then] | |
1265 | |
1266 conditions = [ | |
1267 cond[:2] for cond in conditions if cond[1] | |
1268 ] + [ | |
1269 if_not_os(parse_condition(cond), cond[2]) | |
1270 for cond in conditions if len(cond) == 3 | |
1271 ] | |
1272 | |
1273 if 'variables' in value: | |
1274 conditions.append(if_not_os(None, {'variables': value.pop('variables')})) | |
1275 conditions.sort() | |
1276 | |
1277 value = value.copy() | |
1278 value['conditions'] = conditions | |
1279 return value | |
1280 | |
1281 | |
1282 def load_isolate_as_config(isolate_dir, value, file_comment): | |
1283 """Parses one .isolate file and returns a Configs() instance. | |
1284 | |
1285 |value| is the loaded dictionary that was defined in the gyp file. | |
1286 | |
1287 The expected format is strict, anything diverting from the format below will | |
1288 throw an assert: | |
1289 { | |
1290 'includes': [ | |
1291 'foo.isolate', | |
1292 ], | |
1293 'conditions': [ | |
1294 ['OS=="vms" and foo=42', { | |
1295 'variables': { | |
1296 'command': [ | |
1297 ... | |
1298 ], | |
1299 'isolate_dependency_tracked': [ | |
1300 ... | |
1301 ], | |
1302 'isolate_dependency_untracked': [ | |
1303 ... | |
1304 ], | |
1305 'read_only': False, | |
1306 }, | |
1307 }], | |
1308 ... | |
1309 ], | |
1310 } | |
1311 """ | |
1312 value = convert_old_to_new_format(value) | |
1313 | |
1314 variables_and_values = {} | |
1315 verify_root(value, variables_and_values) | |
1316 if variables_and_values: | |
1317 config_variables, config_values = zip( | |
1318 *sorted(variables_and_values.iteritems())) | |
1319 all_configs = list(itertools.product(*config_values)) | |
1320 else: | |
1321 config_variables = None | |
1322 all_configs = [] | |
1323 | |
1324 isolate = Configs(file_comment) | |
1325 | |
1326 # Add configuration-specific variables. | |
1327 for expr, then in value.get('conditions', []): | |
1328 configs = match_configs(expr, config_variables, all_configs) | |
1329 isolate.merge_dependencies(then['variables'], config_variables, configs) | |
1330 | |
1331 # Load the includes. | |
1332 for include in value.get('includes', []): | |
1333 if os.path.isabs(include): | |
1334 raise ExecutionError( | |
1335 'Failed to load configuration; absolute include path \'%s\'' % | |
1336 include) | |
1337 included_isolate = os.path.normpath(os.path.join(isolate_dir, include)) | |
1338 with open(included_isolate, 'r') as f: | |
1339 included_isolate = load_isolate_as_config( | |
1340 os.path.dirname(included_isolate), | |
1341 eval_content(f.read()), | |
1342 None) | |
1343 isolate = union(isolate, included_isolate) | |
1344 | |
1345 return isolate | |
1346 | |
1347 | |
1348 def load_isolate_for_config(isolate_dir, content, variables): | |
1349 """Loads the .isolate file and returns the information unprocessed but | |
1350 filtered for the specific OS. | |
1351 | |
1352 Returns the command, dependencies and read_only flag. The dependencies are | |
1353 fixed to use os.path.sep. | |
1354 """ | |
1355 # Load the .isolate file, process its conditions, retrieve the command and | |
1356 # dependencies. | |
1357 isolate = load_isolate_as_config(isolate_dir, eval_content(content), None) | |
1358 try: | |
1359 config_name = tuple(variables[var] for var in isolate.config_variables) | |
1360 except KeyError: | |
1361 raise ExecutionError( | |
1362 'These configuration variables were missing from the command line: %s' % | |
1363 ', '.join(sorted(set(isolate.config_variables) - set(variables)))) | |
1364 config = isolate.by_config.get(config_name) | |
1365 if not config: | |
1366 raise ExecutionError( | |
1367 'Failed to load configuration for variable \'%s\' for config(s) \'%s\'' | |
1368 '\nAvailable configs: %s' % | |
1369 (', '.join(isolate.config_variables), | |
1370 ', '.join(config_name), | |
1371 ', '.join(str(s) for s in isolate.by_config))) | |
1372 # Merge tracked and untracked variables, isolate.py doesn't care about the | |
1373 # trackability of the variables, only the build tool does. | |
1374 dependencies = [ | |
1375 f.replace('/', os.path.sep) for f in config.tracked + config.untracked | |
1376 ] | |
1377 touched = [f.replace('/', os.path.sep) for f in config.touched] | |
1378 return config.command, dependencies, touched, config.read_only | |
1379 | |
1380 | |
1381 def save_isolated(isolated, data): | |
1382 """Writes one or multiple .isolated files. | |
1383 | |
1384 Note: this reference implementation does not create child .isolated file so it | |
1385 always returns an empty list. | |
1386 | |
1387 Returns the list of child isolated files that are included by |isolated|. | |
1388 """ | |
1389 trace_inputs.write_json(isolated, data, True) | |
1390 return [] | |
1391 | |
1392 | |
1393 def chromium_save_isolated(isolated, data, variables, algo): | |
1394 """Writes one or many .isolated files. | |
1395 | |
1396 This slightly increases the cold cache cost but greatly reduce the warm cache | |
1397 cost by splitting low-churn files off the master .isolated file. It also | |
1398 reduces overall isolateserver memcache consumption. | |
1399 """ | |
1400 slaves = [] | |
1401 | |
1402 def extract_into_included_isolated(prefix): | |
1403 new_slave = { | |
1404 'algo': data['algo'], | |
1405 'files': {}, | |
1406 'os': data['os'], | |
1407 'version': data['version'], | |
1408 } | |
1409 for f in data['files'].keys(): | |
1410 if f.startswith(prefix): | |
1411 new_slave['files'][f] = data['files'].pop(f) | |
1412 if new_slave['files']: | |
1413 slaves.append(new_slave) | |
1414 | |
1415 # Split test/data/ in its own .isolated file. | |
1416 extract_into_included_isolated(os.path.join('test', 'data', '')) | |
1417 | |
1418 # Split everything out of PRODUCT_DIR in its own .isolated file. | |
1419 if variables.get('PRODUCT_DIR'): | |
1420 extract_into_included_isolated(variables['PRODUCT_DIR']) | |
1421 | |
1422 files = [] | |
1423 for index, f in enumerate(slaves): | |
1424 slavepath = isolated[:-len('.isolated')] + '.%d.isolated' % index | |
1425 trace_inputs.write_json(slavepath, f, True) | |
1426 data.setdefault('includes', []).append( | |
1427 isolateserver.hash_file(slavepath, algo)) | |
1428 files.append(os.path.basename(slavepath)) | |
1429 | |
1430 files.extend(save_isolated(isolated, data)) | |
1431 return files | |
1432 | |
1433 | |
1434 class Flattenable(object): | |
1435 """Represents data that can be represented as a json file.""" | |
1436 MEMBERS = () | |
1437 | |
1438 def flatten(self): | |
1439 """Returns a json-serializable version of itself. | |
1440 | |
1441 Skips None entries. | |
1442 """ | |
1443 items = ((member, getattr(self, member)) for member in self.MEMBERS) | |
1444 return dict((member, value) for member, value in items if value is not None) | |
1445 | |
1446 @classmethod | |
1447 def load(cls, data, *args, **kwargs): | |
1448 """Loads a flattened version.""" | |
1449 data = data.copy() | |
1450 out = cls(*args, **kwargs) | |
1451 for member in out.MEMBERS: | |
1452 if member in data: | |
1453 # Access to a protected member XXX of a client class | |
1454 # pylint: disable=W0212 | |
1455 out._load_member(member, data.pop(member)) | |
1456 if data: | |
1457 raise ValueError( | |
1458 'Found unexpected entry %s while constructing an object %s' % | |
1459 (data, cls.__name__), data, cls.__name__) | |
1460 return out | |
1461 | |
1462 def _load_member(self, member, value): | |
1463 """Loads a member into self.""" | |
1464 setattr(self, member, value) | |
1465 | |
1466 @classmethod | |
1467 def load_file(cls, filename, *args, **kwargs): | |
1468 """Loads the data from a file or return an empty instance.""" | |
1469 try: | |
1470 out = cls.load(trace_inputs.read_json(filename), *args, **kwargs) | |
1471 logging.debug('Loaded %s(%s)', cls.__name__, filename) | |
1472 except (IOError, ValueError) as e: | |
1473 # On failure, loads the default instance. | |
1474 out = cls(*args, **kwargs) | |
1475 logging.warn('Failed to load %s: %s', filename, e) | |
1476 return out | |
1477 | |
1478 | |
1479 class SavedState(Flattenable): | |
1480 """Describes the content of a .state file. | |
1481 | |
1482 This file caches the items calculated by this script and is used to increase | |
1483 the performance of the script. This file is not loaded by run_isolated.py. | |
1484 This file can always be safely removed. | |
1485 | |
1486 It is important to note that the 'files' dict keys are using native OS path | |
1487 separator instead of '/' used in .isolate file. | |
1488 """ | |
1489 MEMBERS = ( | |
1490 # Algorithm used to generate the hash. The only supported value is at the | |
1491 # time of writting 'sha-1'. | |
1492 'algo', | |
1493 # Cache of the processed command. This value is saved because .isolated | |
1494 # files are never loaded by isolate.py so it's the only way to load the | |
1495 # command safely. | |
1496 'command', | |
1497 # Cache of the files found so the next run can skip hash calculation. | |
1498 'files', | |
1499 # Path of the original .isolate file. Relative path to isolated_basedir. | |
1500 'isolate_file', | |
1501 # List of included .isolated files. Used to support/remember 'slave' | |
1502 # .isolated files. Relative path to isolated_basedir. | |
1503 'child_isolated_files', | |
1504 # If the generated directory tree should be read-only. | |
1505 'read_only', | |
1506 # Relative cwd to use to start the command. | |
1507 'relative_cwd', | |
1508 # GYP variables used to generate the .isolated file. Variables are saved so | |
1509 # a user can use isolate.py after building and the GYP variables are still | |
1510 # defined. | |
1511 'variables', | |
1512 # Version of the file format in format 'major.minor'. Any non-breaking | |
1513 # change must update minor. Any breaking change must update major. | |
1514 'version', | |
1515 ) | |
1516 | |
1517 def __init__(self, isolated_basedir): | |
1518 """Creates an empty SavedState. | |
1519 | |
1520 |isolated_basedir| is the directory where the .isolated and .isolated.state | |
1521 files are saved. | |
1522 """ | |
1523 super(SavedState, self).__init__() | |
1524 assert os.path.isabs(isolated_basedir), isolated_basedir | |
1525 assert os.path.isdir(isolated_basedir), isolated_basedir | |
1526 self.isolated_basedir = isolated_basedir | |
1527 | |
1528 # The default algorithm used. | |
1529 self.algo = isolateserver.SUPPORTED_ALGOS['sha-1'] | |
1530 self.command = [] | |
1531 self.files = {} | |
1532 self.isolate_file = None | |
1533 self.child_isolated_files = [] | |
1534 self.read_only = None | |
1535 self.relative_cwd = None | |
1536 self.variables = {'OS': get_flavor()} | |
1537 # The current version. | |
1538 self.version = '1.0' | |
1539 | |
1540 def update(self, isolate_file, variables): | |
1541 """Updates the saved state with new data to keep GYP variables and internal | |
1542 reference to the original .isolate file. | |
1543 """ | |
1544 assert os.path.isabs(isolate_file) | |
1545 # Convert back to a relative path. On Windows, if the isolate and | |
1546 # isolated files are on different drives, isolate_file will stay an absolute | |
1547 # path. | |
1548 isolate_file = safe_relpath(isolate_file, self.isolated_basedir) | |
1549 | |
1550 # The same .isolate file should always be used to generate the .isolated and | |
1551 # .isolated.state. | |
1552 assert isolate_file == self.isolate_file or not self.isolate_file, ( | |
1553 isolate_file, self.isolate_file) | |
1554 self.isolate_file = isolate_file | |
1555 self.variables.update(variables) | |
1556 | |
1557 def update_isolated(self, command, infiles, touched, read_only, relative_cwd): | |
1558 """Updates the saved state with data necessary to generate a .isolated file. | |
1559 | |
1560 The new files in |infiles| are added to self.files dict but their hash is | |
1561 not calculated here. | |
1562 """ | |
1563 self.command = command | |
1564 # Add new files. | |
1565 for f in infiles: | |
1566 self.files.setdefault(f, {}) | |
1567 for f in touched: | |
1568 self.files.setdefault(f, {})['T'] = True | |
1569 # Prune extraneous files that are not a dependency anymore. | |
1570 for f in set(self.files).difference(set(infiles).union(touched)): | |
1571 del self.files[f] | |
1572 if read_only is not None: | |
1573 self.read_only = read_only | |
1574 self.relative_cwd = relative_cwd | |
1575 | |
1576 def to_isolated(self): | |
1577 """Creates a .isolated dictionary out of the saved state. | |
1578 | |
1579 https://code.google.com/p/swarming/wiki/IsolatedDesign | |
1580 """ | |
1581 def strip(data): | |
1582 """Returns a 'files' entry with only the whitelisted keys.""" | |
1583 return dict((k, data[k]) for k in ('h', 'l', 'm', 's') if k in data) | |
1584 | |
1585 out = { | |
1586 'algo': isolateserver.SUPPORTED_ALGOS_REVERSE[self.algo], | |
1587 'files': dict( | |
1588 (filepath, strip(data)) for filepath, data in self.files.iteritems()), | |
1589 'os': self.variables['OS'], | |
1590 'version': self.version, | |
1591 } | |
1592 if self.command: | |
1593 out['command'] = self.command | |
1594 if self.read_only is not None: | |
1595 out['read_only'] = self.read_only | |
1596 if self.relative_cwd: | |
1597 out['relative_cwd'] = self.relative_cwd | |
1598 return out | |
1599 | |
1600 @property | |
1601 def isolate_filepath(self): | |
1602 """Returns the absolute path of self.isolate_file.""" | |
1603 return os.path.normpath( | |
1604 os.path.join(self.isolated_basedir, self.isolate_file)) | |
1605 | |
1606 # Arguments number differs from overridden method | |
1607 @classmethod | |
1608 def load(cls, data, isolated_basedir): # pylint: disable=W0221 | |
1609 """Special case loading to disallow different OS. | |
1610 | |
1611 It is not possible to load a .isolated.state files from a different OS, this | |
1612 file is saved in OS-specific format. | |
1613 """ | |
1614 out = super(SavedState, cls).load(data, isolated_basedir) | |
1615 if 'os' in data: | |
1616 out.variables['OS'] = data['os'] | |
1617 | |
1618 # Converts human readable form back into the proper class type. | |
1619 algo = data.get('algo', 'sha-1') | |
1620 if not algo in isolateserver.SUPPORTED_ALGOS: | |
1621 raise isolateserver.ConfigError('Unknown algo \'%s\'' % out.algo) | |
1622 out.algo = isolateserver.SUPPORTED_ALGOS[algo] | |
1623 | |
1624 # For example, 1.1 is guaranteed to be backward compatible with 1.0 code. | |
1625 if not re.match(r'^(\d+)\.(\d+)$', out.version): | |
1626 raise isolateserver.ConfigError('Unknown version \'%s\'' % out.version) | |
1627 if out.version.split('.', 1)[0] != '1': | |
1628 raise isolateserver.ConfigError( | |
1629 'Unsupported version \'%s\'' % out.version) | |
1630 | |
1631 # The .isolate file must be valid. It could be absolute on Windows if the | |
1632 # drive containing the .isolate and the drive containing the .isolated files | |
1633 # differ. | |
1634 assert not os.path.isabs(out.isolate_file) or sys.platform == 'win32' | |
1635 assert os.path.isfile(out.isolate_filepath), out.isolate_filepath | |
1636 return out | |
1637 | |
1638 def flatten(self): | |
1639 """Makes sure 'algo' is in human readable form.""" | |
1640 out = super(SavedState, self).flatten() | |
1641 out['algo'] = isolateserver.SUPPORTED_ALGOS_REVERSE[out['algo']] | |
1642 return out | |
1643 | |
1644 def __str__(self): | |
1645 out = '%s(\n' % self.__class__.__name__ | |
1646 out += ' command: %s\n' % self.command | |
1647 out += ' files: %d\n' % len(self.files) | |
1648 out += ' isolate_file: %s\n' % self.isolate_file | |
1649 out += ' read_only: %s\n' % self.read_only | |
1650 out += ' relative_cwd: %s\n' % self.relative_cwd | |
1651 out += ' child_isolated_files: %s\n' % self.child_isolated_files | |
1652 out += ' variables: %s' % ''.join( | |
1653 '\n %s=%s' % (k, self.variables[k]) for k in sorted(self.variables)) | |
1654 out += ')' | |
1655 return out | |
1656 | |
1657 | |
1658 class CompleteState(object): | |
1659 """Contains all the state to run the task at hand.""" | |
1660 def __init__(self, isolated_filepath, saved_state): | |
1661 super(CompleteState, self).__init__() | |
1662 assert isolated_filepath is None or os.path.isabs(isolated_filepath) | |
1663 self.isolated_filepath = isolated_filepath | |
1664 # Contains the data to ease developer's use-case but that is not strictly | |
1665 # necessary. | |
1666 self.saved_state = saved_state | |
1667 | |
1668 @classmethod | |
1669 def load_files(cls, isolated_filepath): | |
1670 """Loads state from disk.""" | |
1671 assert os.path.isabs(isolated_filepath), isolated_filepath | |
1672 isolated_basedir = os.path.dirname(isolated_filepath) | |
1673 return cls( | |
1674 isolated_filepath, | |
1675 SavedState.load_file( | |
1676 isolatedfile_to_state(isolated_filepath), isolated_basedir)) | |
1677 | |
1678 def load_isolate(self, cwd, isolate_file, variables, ignore_broken_items): | |
1679 """Updates self.isolated and self.saved_state with information loaded from a | |
1680 .isolate file. | |
1681 | |
1682 Processes the loaded data, deduce root_dir, relative_cwd. | |
1683 """ | |
1684 # Make sure to not depend on os.getcwd(). | |
1685 assert os.path.isabs(isolate_file), isolate_file | |
1686 isolate_file = file_path.get_native_path_case(isolate_file) | |
1687 logging.info( | |
1688 'CompleteState.load_isolate(%s, %s, %s, %s)', | |
1689 cwd, isolate_file, variables, ignore_broken_items) | |
1690 relative_base_dir = os.path.dirname(isolate_file) | |
1691 | |
1692 # Processes the variables and update the saved state. | |
1693 variables = process_variables(cwd, variables, relative_base_dir) | |
1694 self.saved_state.update(isolate_file, variables) | |
1695 variables = self.saved_state.variables | |
1696 | |
1697 with open(isolate_file, 'r') as f: | |
1698 # At that point, variables are not replaced yet in command and infiles. | |
1699 # infiles may contain directory entries and is in posix style. | |
1700 command, infiles, touched, read_only = load_isolate_for_config( | |
1701 os.path.dirname(isolate_file), f.read(), variables) | |
1702 command = [eval_variables(i, variables) for i in command] | |
1703 infiles = [eval_variables(f, variables) for f in infiles] | |
1704 touched = [eval_variables(f, variables) for f in touched] | |
1705 # root_dir is automatically determined by the deepest root accessed with the | |
1706 # form '../../foo/bar'. Note that path variables must be taken in account | |
1707 # too, add them as if they were input files. | |
1708 path_variables = [variables[v] for v in PATH_VARIABLES if v in variables] | |
1709 root_dir = determine_root_dir( | |
1710 relative_base_dir, infiles + touched + path_variables) | |
1711 # The relative directory is automatically determined by the relative path | |
1712 # between root_dir and the directory containing the .isolate file, | |
1713 # isolate_base_dir. | |
1714 relative_cwd = os.path.relpath(relative_base_dir, root_dir) | |
1715 # Now that we know where the root is, check that the PATH_VARIABLES point | |
1716 # inside it. | |
1717 for i in PATH_VARIABLES: | |
1718 if i in variables: | |
1719 if not path_starts_with( | |
1720 root_dir, os.path.join(relative_base_dir, variables[i])): | |
1721 raise isolateserver.MappingError( | |
1722 'Path variable %s=%r points outside the inferred root directory' | |
1723 ' %s' % (i, variables[i], root_dir)) | |
1724 # Normalize the files based to root_dir. It is important to keep the | |
1725 # trailing os.path.sep at that step. | |
1726 infiles = [ | |
1727 relpath(normpath(os.path.join(relative_base_dir, f)), root_dir) | |
1728 for f in infiles | |
1729 ] | |
1730 touched = [ | |
1731 relpath(normpath(os.path.join(relative_base_dir, f)), root_dir) | |
1732 for f in touched | |
1733 ] | |
1734 follow_symlinks = variables['OS'] != 'win' | |
1735 # Expand the directories by listing each file inside. Up to now, trailing | |
1736 # os.path.sep must be kept. Do not expand 'touched'. | |
1737 infiles = expand_directories_and_symlinks( | |
1738 root_dir, | |
1739 infiles, | |
1740 lambda x: re.match(r'.*\.(git|svn|pyc)$', x), | |
1741 follow_symlinks, | |
1742 ignore_broken_items) | |
1743 | |
1744 # If we ignore broken items then remove any missing touched items. | |
1745 if ignore_broken_items: | |
1746 original_touched_count = len(touched) | |
1747 touched = [touch for touch in touched if os.path.exists(touch)] | |
1748 | |
1749 if len(touched) != original_touched_count: | |
1750 logging.info('Removed %d invalid touched entries', | |
1751 len(touched) - original_touched_count) | |
1752 | |
1753 # Finally, update the new data to be able to generate the foo.isolated file, | |
1754 # the file that is used by run_isolated.py. | |
1755 self.saved_state.update_isolated( | |
1756 command, infiles, touched, read_only, relative_cwd) | |
1757 logging.debug(self) | |
1758 | |
1759 def process_inputs(self, subdir): | |
1760 """Updates self.saved_state.files with the files' mode and hash. | |
1761 | |
1762 If |subdir| is specified, filters to a subdirectory. The resulting .isolated | |
1763 file is tainted. | |
1764 | |
1765 See process_input() for more information. | |
1766 """ | |
1767 for infile in sorted(self.saved_state.files): | |
1768 if subdir and not infile.startswith(subdir): | |
1769 self.saved_state.files.pop(infile) | |
1770 else: | |
1771 filepath = os.path.join(self.root_dir, infile) | |
1772 self.saved_state.files[infile] = process_input( | |
1773 filepath, | |
1774 self.saved_state.files[infile], | |
1775 self.saved_state.read_only, | |
1776 self.saved_state.variables['OS'], | |
1777 self.saved_state.algo) | |
1778 | |
1779 def save_files(self): | |
1780 """Saves self.saved_state and creates a .isolated file.""" | |
1781 logging.debug('Dumping to %s' % self.isolated_filepath) | |
1782 self.saved_state.child_isolated_files = chromium_save_isolated( | |
1783 self.isolated_filepath, | |
1784 self.saved_state.to_isolated(), | |
1785 self.saved_state.variables, | |
1786 self.saved_state.algo) | |
1787 total_bytes = sum( | |
1788 i.get('s', 0) for i in self.saved_state.files.itervalues()) | |
1789 if total_bytes: | |
1790 # TODO(maruel): Stats are missing the .isolated files. | |
1791 logging.debug('Total size: %d bytes' % total_bytes) | |
1792 saved_state_file = isolatedfile_to_state(self.isolated_filepath) | |
1793 logging.debug('Dumping to %s' % saved_state_file) | |
1794 trace_inputs.write_json(saved_state_file, self.saved_state.flatten(), True) | |
1795 | |
1796 @property | |
1797 def root_dir(self): | |
1798 """Returns the absolute path of the root_dir to reference the .isolate file | |
1799 via relative_cwd. | |
1800 | |
1801 So that join(root_dir, relative_cwd, basename(isolate_file)) is equivalent | |
1802 to isolate_filepath. | |
1803 """ | |
1804 if not self.saved_state.isolate_file: | |
1805 raise ExecutionError('Please specify --isolate') | |
1806 isolate_dir = os.path.dirname(self.saved_state.isolate_filepath) | |
1807 # Special case '.'. | |
1808 if self.saved_state.relative_cwd == '.': | |
1809 root_dir = isolate_dir | |
1810 else: | |
1811 if not isolate_dir.endswith(self.saved_state.relative_cwd): | |
1812 raise ExecutionError( | |
1813 ('Make sure the .isolate file is in the directory that will be ' | |
1814 'used as the relative directory. It is currently in %s and should ' | |
1815 'be in %s') % (isolate_dir, self.saved_state.relative_cwd)) | |
1816 # Walk back back to the root directory. | |
1817 root_dir = isolate_dir[:-(len(self.saved_state.relative_cwd) + 1)] | |
1818 return file_path.get_native_path_case(root_dir) | |
1819 | |
1820 @property | |
1821 def resultdir(self): | |
1822 """Returns the absolute path containing the .isolated file. | |
1823 | |
1824 It is usually equivalent to the variable PRODUCT_DIR. Uses the .isolated | |
1825 path as the value. | |
1826 """ | |
1827 return os.path.dirname(self.isolated_filepath) | |
1828 | |
1829 def __str__(self): | |
1830 def indent(data, indent_length): | |
1831 """Indents text.""" | |
1832 spacing = ' ' * indent_length | |
1833 return ''.join(spacing + l for l in str(data).splitlines(True)) | |
1834 | |
1835 out = '%s(\n' % self.__class__.__name__ | |
1836 out += ' root_dir: %s\n' % self.root_dir | |
1837 out += ' saved_state: %s)' % indent(self.saved_state, 2) | |
1838 return out | |
1839 | |
1840 | |
1841 def load_complete_state(options, cwd, subdir, skip_update): | |
1842 """Loads a CompleteState. | |
1843 | |
1844 This includes data from .isolate and .isolated.state files. Never reads the | |
1845 .isolated file. | |
1846 | |
1847 Arguments: | |
1848 options: Options instance generated with OptionParserIsolate. For either | |
1849 options.isolate and options.isolated, if the value is set, it is an | |
1850 absolute path. | |
1851 cwd: base directory to be used when loading the .isolate file. | |
1852 subdir: optional argument to only process file in the subdirectory, relative | |
1853 to CompleteState.root_dir. | |
1854 skip_update: Skip trying to load the .isolate file and processing the | |
1855 dependencies. It is useful when not needed, like when tracing. | |
1856 """ | |
1857 assert not options.isolate or os.path.isabs(options.isolate) | |
1858 assert not options.isolated or os.path.isabs(options.isolated) | |
1859 cwd = file_path.get_native_path_case(unicode(cwd)) | |
1860 if options.isolated: | |
1861 # Load the previous state if it was present. Namely, "foo.isolated.state". | |
1862 # Note: this call doesn't load the .isolate file. | |
1863 complete_state = CompleteState.load_files(options.isolated) | |
1864 else: | |
1865 # Constructs a dummy object that cannot be saved. Useful for temporary | |
1866 # commands like 'run'. | |
1867 complete_state = CompleteState(None, SavedState()) | |
1868 | |
1869 if not options.isolate: | |
1870 if not complete_state.saved_state.isolate_file: | |
1871 if not skip_update: | |
1872 raise ExecutionError('A .isolate file is required.') | |
1873 isolate = None | |
1874 else: | |
1875 isolate = complete_state.saved_state.isolate_filepath | |
1876 else: | |
1877 isolate = options.isolate | |
1878 if complete_state.saved_state.isolate_file: | |
1879 rel_isolate = safe_relpath( | |
1880 options.isolate, complete_state.saved_state.isolated_basedir) | |
1881 if rel_isolate != complete_state.saved_state.isolate_file: | |
1882 raise ExecutionError( | |
1883 '%s and %s do not match.' % ( | |
1884 options.isolate, complete_state.saved_state.isolate_file)) | |
1885 | |
1886 if not skip_update: | |
1887 # Then load the .isolate and expands directories. | |
1888 complete_state.load_isolate( | |
1889 cwd, isolate, options.variables, options.ignore_broken_items) | |
1890 | |
1891 # Regenerate complete_state.saved_state.files. | |
1892 if subdir: | |
1893 subdir = unicode(subdir) | |
1894 subdir = eval_variables(subdir, complete_state.saved_state.variables) | |
1895 subdir = subdir.replace('/', os.path.sep) | |
1896 | |
1897 if not skip_update: | |
1898 complete_state.process_inputs(subdir) | |
1899 return complete_state | |
1900 | |
1901 | |
1902 def read_trace_as_isolate_dict(complete_state, trace_blacklist): | |
1903 """Reads a trace and returns the .isolate dictionary. | |
1904 | |
1905 Returns exceptions during the log parsing so it can be re-raised. | |
1906 """ | |
1907 api = trace_inputs.get_api() | |
1908 logfile = complete_state.isolated_filepath + '.log' | |
1909 if not os.path.isfile(logfile): | |
1910 raise ExecutionError( | |
1911 'No log file \'%s\' to read, did you forget to \'trace\'?' % logfile) | |
1912 try: | |
1913 data = api.parse_log(logfile, trace_blacklist, None) | |
1914 exceptions = [i['exception'] for i in data if 'exception' in i] | |
1915 results = (i['results'] for i in data if 'results' in i) | |
1916 results_stripped = (i.strip_root(complete_state.root_dir) for i in results) | |
1917 files = set(sum((result.existent for result in results_stripped), [])) | |
1918 tracked, touched = split_touched(files) | |
1919 value = generate_isolate( | |
1920 tracked, | |
1921 [], | |
1922 touched, | |
1923 complete_state.root_dir, | |
1924 complete_state.saved_state.variables, | |
1925 complete_state.saved_state.relative_cwd, | |
1926 trace_blacklist) | |
1927 return value, exceptions | |
1928 except trace_inputs.TracingFailure, e: | |
1929 raise ExecutionError( | |
1930 'Reading traces failed for: %s\n%s' % | |
1931 (' '.join(complete_state.saved_state.command), str(e))) | |
1932 | |
1933 | |
1934 def print_all(comment, data, stream): | |
1935 """Prints a complete .isolate file and its top-level file comment into a | |
1936 stream. | |
1937 """ | |
1938 if comment: | |
1939 stream.write(comment) | |
1940 pretty_print(data, stream) | |
1941 | |
1942 | |
1943 def merge(complete_state, trace_blacklist): | |
1944 """Reads a trace and merges it back into the source .isolate file.""" | |
1945 value, exceptions = read_trace_as_isolate_dict( | |
1946 complete_state, trace_blacklist) | |
1947 | |
1948 # Now take that data and union it into the original .isolate file. | |
1949 with open(complete_state.saved_state.isolate_filepath, 'r') as f: | |
1950 prev_content = f.read() | |
1951 isolate_dir = os.path.dirname(complete_state.saved_state.isolate_filepath) | |
1952 prev_config = load_isolate_as_config( | |
1953 isolate_dir, | |
1954 eval_content(prev_content), | |
1955 extract_comment(prev_content)) | |
1956 new_config = load_isolate_as_config(isolate_dir, value, '') | |
1957 config = union(prev_config, new_config) | |
1958 data = config.make_isolate_file() | |
1959 print('Updating %s' % complete_state.saved_state.isolate_file) | |
1960 with open(complete_state.saved_state.isolate_filepath, 'wb') as f: | |
1961 print_all(config.file_comment, data, f) | |
1962 if exceptions: | |
1963 # It got an exception, raise the first one. | |
1964 raise \ | |
1965 exceptions[0][0], \ | |
1966 exceptions[0][1], \ | |
1967 exceptions[0][2] | |
1968 | |
1969 | |
1970 ### Commands. | |
1971 | |
1972 | |
1973 def CMDarchive(parser, args): | |
1974 """Creates a .isolated file and uploads the tree to an isolate server. | |
1975 | |
1976 All the files listed in the .isolated file are put in the isolate server | |
1977 cache via isolateserver.py. | |
1978 """ | |
1979 parser.add_option('--subdir', help='Filters to a subdirectory') | |
1980 options, args = parser.parse_args(args) | |
1981 if args: | |
1982 parser.error('Unsupported argument: %s' % args) | |
1983 | |
1984 with tools.Profiler('GenerateHashtable'): | |
1985 success = False | |
1986 try: | |
1987 complete_state = load_complete_state( | |
1988 options, os.getcwd(), options.subdir, False) | |
1989 if not options.outdir: | |
1990 options.outdir = os.path.join( | |
1991 os.path.dirname(complete_state.isolated_filepath), 'hashtable') | |
1992 # Make sure that complete_state isn't modified until save_files() is | |
1993 # called, because any changes made to it here will propagate to the files | |
1994 # created (which is probably not intended). | |
1995 complete_state.save_files() | |
1996 | |
1997 infiles = complete_state.saved_state.files | |
1998 # Add all the .isolated files. | |
1999 isolated_hash = [] | |
2000 isolated_files = [ | |
2001 options.isolated, | |
2002 ] + complete_state.saved_state.child_isolated_files | |
2003 for item in isolated_files: | |
2004 item_path = os.path.join( | |
2005 os.path.dirname(complete_state.isolated_filepath), item) | |
2006 # Do not use isolateserver.hash_file() here because the file is | |
2007 # likely smallish (under 500kb) and its file size is needed. | |
2008 with open(item_path, 'rb') as f: | |
2009 content = f.read() | |
2010 isolated_hash.append( | |
2011 complete_state.saved_state.algo(content).hexdigest()) | |
2012 isolated_metadata = { | |
2013 'h': isolated_hash[-1], | |
2014 's': len(content), | |
2015 'priority': '0' | |
2016 } | |
2017 infiles[item_path] = isolated_metadata | |
2018 | |
2019 logging.info('Creating content addressed object store with %d item', | |
2020 len(infiles)) | |
2021 | |
2022 if is_url(options.outdir): | |
2023 isolateserver.upload_tree( | |
2024 base_url=options.outdir, | |
2025 indir=complete_state.root_dir, | |
2026 infiles=infiles, | |
2027 namespace='default-gzip') | |
2028 else: | |
2029 recreate_tree( | |
2030 outdir=options.outdir, | |
2031 indir=complete_state.root_dir, | |
2032 infiles=infiles, | |
2033 action=run_isolated.HARDLINK_WITH_FALLBACK, | |
2034 as_hash=True) | |
2035 success = True | |
2036 print('%s %s' % (isolated_hash[0], os.path.basename(options.isolated))) | |
2037 finally: | |
2038 # If the command failed, delete the .isolated file if it exists. This is | |
2039 # important so no stale swarm job is executed. | |
2040 if not success and os.path.isfile(options.isolated): | |
2041 os.remove(options.isolated) | |
2042 return not success | |
2043 | |
2044 | |
2045 def CMDcheck(parser, args): | |
2046 """Checks that all the inputs are present and generates .isolated.""" | |
2047 parser.add_option('--subdir', help='Filters to a subdirectory') | |
2048 options, args = parser.parse_args(args) | |
2049 if args: | |
2050 parser.error('Unsupported argument: %s' % args) | |
2051 | |
2052 complete_state = load_complete_state( | |
2053 options, os.getcwd(), options.subdir, False) | |
2054 | |
2055 # Nothing is done specifically. Just store the result and state. | |
2056 complete_state.save_files() | |
2057 return 0 | |
2058 | |
2059 | |
2060 CMDhashtable = CMDarchive | |
2061 | |
2062 | |
2063 def CMDmerge(parser, args): | |
2064 """Reads and merges the data from the trace back into the original .isolate. | |
2065 | |
2066 Ignores --outdir. | |
2067 """ | |
2068 parser.require_isolated = False | |
2069 add_trace_option(parser) | |
2070 options, args = parser.parse_args(args) | |
2071 if args: | |
2072 parser.error('Unsupported argument: %s' % args) | |
2073 | |
2074 complete_state = load_complete_state(options, os.getcwd(), None, False) | |
2075 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist) | |
2076 merge(complete_state, blacklist) | |
2077 return 0 | |
2078 | |
2079 | |
2080 def CMDread(parser, args): | |
2081 """Reads the trace file generated with command 'trace'. | |
2082 | |
2083 Ignores --outdir. | |
2084 """ | |
2085 parser.require_isolated = False | |
2086 add_trace_option(parser) | |
2087 parser.add_option( | |
2088 '--skip-refresh', action='store_true', | |
2089 help='Skip reading .isolate file and do not refresh the hash of ' | |
2090 'dependencies') | |
2091 parser.add_option( | |
2092 '-m', '--merge', action='store_true', | |
2093 help='merge the results back in the .isolate file instead of printing') | |
2094 options, args = parser.parse_args(args) | |
2095 if args: | |
2096 parser.error('Unsupported argument: %s' % args) | |
2097 | |
2098 complete_state = load_complete_state( | |
2099 options, os.getcwd(), None, options.skip_refresh) | |
2100 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist) | |
2101 value, exceptions = read_trace_as_isolate_dict(complete_state, blacklist) | |
2102 if options.merge: | |
2103 merge(complete_state, blacklist) | |
2104 else: | |
2105 pretty_print(value, sys.stdout) | |
2106 | |
2107 if exceptions: | |
2108 # It got an exception, raise the first one. | |
2109 raise \ | |
2110 exceptions[0][0], \ | |
2111 exceptions[0][1], \ | |
2112 exceptions[0][2] | |
2113 return 0 | |
2114 | |
2115 | |
2116 def CMDremap(parser, args): | |
2117 """Creates a directory with all the dependencies mapped into it. | |
2118 | |
2119 Useful to test manually why a test is failing. The target executable is not | |
2120 run. | |
2121 """ | |
2122 parser.require_isolated = False | |
2123 options, args = parser.parse_args(args) | |
2124 if args: | |
2125 parser.error('Unsupported argument: %s' % args) | |
2126 complete_state = load_complete_state(options, os.getcwd(), None, False) | |
2127 | |
2128 if not options.outdir: | |
2129 options.outdir = run_isolated.make_temp_dir( | |
2130 'isolate', complete_state.root_dir) | |
2131 else: | |
2132 if is_url(options.outdir): | |
2133 parser.error('Can\'t use url for --outdir with mode remap.') | |
2134 if not os.path.isdir(options.outdir): | |
2135 os.makedirs(options.outdir) | |
2136 print('Remapping into %s' % options.outdir) | |
2137 if len(os.listdir(options.outdir)): | |
2138 raise ExecutionError('Can\'t remap in a non-empty directory') | |
2139 recreate_tree( | |
2140 outdir=options.outdir, | |
2141 indir=complete_state.root_dir, | |
2142 infiles=complete_state.saved_state.files, | |
2143 action=run_isolated.HARDLINK_WITH_FALLBACK, | |
2144 as_hash=False) | |
2145 if complete_state.saved_state.read_only: | |
2146 run_isolated.make_writable(options.outdir, True) | |
2147 | |
2148 if complete_state.isolated_filepath: | |
2149 complete_state.save_files() | |
2150 return 0 | |
2151 | |
2152 | |
2153 def CMDrewrite(parser, args): | |
2154 """Rewrites a .isolate file into the canonical format.""" | |
2155 parser.require_isolated = False | |
2156 options, args = parser.parse_args(args) | |
2157 if args: | |
2158 parser.error('Unsupported argument: %s' % args) | |
2159 | |
2160 if options.isolated: | |
2161 # Load the previous state if it was present. Namely, "foo.isolated.state". | |
2162 complete_state = CompleteState.load_files(options.isolated) | |
2163 isolate = options.isolate or complete_state.saved_state.isolate_filepath | |
2164 else: | |
2165 isolate = options.isolate | |
2166 if not isolate: | |
2167 parser.error('--isolate is required.') | |
2168 | |
2169 with open(isolate, 'r') as f: | |
2170 content = f.read() | |
2171 config = load_isolate_as_config( | |
2172 os.path.dirname(os.path.abspath(isolate)), | |
2173 eval_content(content), | |
2174 extract_comment(content)) | |
2175 data = config.make_isolate_file() | |
2176 print('Updating %s' % isolate) | |
2177 with open(isolate, 'wb') as f: | |
2178 print_all(config.file_comment, data, f) | |
2179 return 0 | |
2180 | |
2181 | |
2182 @subcommand.usage('-- [extra arguments]') | |
2183 def CMDrun(parser, args): | |
2184 """Runs the test executable in an isolated (temporary) directory. | |
2185 | |
2186 All the dependencies are mapped into the temporary directory and the | |
2187 directory is cleaned up after the target exits. Warning: if --outdir is | |
2188 specified, it is deleted upon exit. | |
2189 | |
2190 Argument processing stops at -- and these arguments are appended to the | |
2191 command line of the target to run. For example, use: | |
2192 isolate.py run --isolated foo.isolated -- --gtest_filter=Foo.Bar | |
2193 """ | |
2194 parser.require_isolated = False | |
2195 parser.add_option( | |
2196 '--skip-refresh', action='store_true', | |
2197 help='Skip reading .isolate file and do not refresh the hash of ' | |
2198 'dependencies') | |
2199 options, args = parser.parse_args(args) | |
2200 if options.outdir and is_url(options.outdir): | |
2201 parser.error('Can\'t use url for --outdir with mode run.') | |
2202 | |
2203 complete_state = load_complete_state( | |
2204 options, os.getcwd(), None, options.skip_refresh) | |
2205 cmd = complete_state.saved_state.command + args | |
2206 if not cmd: | |
2207 raise ExecutionError('No command to run.') | |
2208 | |
2209 cmd = tools.fix_python_path(cmd) | |
2210 try: | |
2211 root_dir = complete_state.root_dir | |
2212 if not options.outdir: | |
2213 if not os.path.isabs(root_dir): | |
2214 root_dir = os.path.join(os.path.dirname(options.isolated), root_dir) | |
2215 options.outdir = run_isolated.make_temp_dir('isolate', root_dir) | |
2216 else: | |
2217 if not os.path.isdir(options.outdir): | |
2218 os.makedirs(options.outdir) | |
2219 recreate_tree( | |
2220 outdir=options.outdir, | |
2221 indir=root_dir, | |
2222 infiles=complete_state.saved_state.files, | |
2223 action=run_isolated.HARDLINK_WITH_FALLBACK, | |
2224 as_hash=False) | |
2225 cwd = os.path.normpath( | |
2226 os.path.join(options.outdir, complete_state.saved_state.relative_cwd)) | |
2227 if not os.path.isdir(cwd): | |
2228 # It can happen when no files are mapped from the directory containing the | |
2229 # .isolate file. But the directory must exist to be the current working | |
2230 # directory. | |
2231 os.makedirs(cwd) | |
2232 if complete_state.saved_state.read_only: | |
2233 run_isolated.make_writable(options.outdir, True) | |
2234 logging.info('Running %s, cwd=%s' % (cmd, cwd)) | |
2235 result = subprocess.call(cmd, cwd=cwd) | |
2236 finally: | |
2237 if options.outdir: | |
2238 run_isolated.rmtree(options.outdir) | |
2239 | |
2240 if complete_state.isolated_filepath: | |
2241 complete_state.save_files() | |
2242 return result | |
2243 | |
2244 | |
2245 @subcommand.usage('-- [extra arguments]') | |
2246 def CMDtrace(parser, args): | |
2247 """Traces the target using trace_inputs.py. | |
2248 | |
2249 It runs the executable without remapping it, and traces all the files it and | |
2250 its child processes access. Then the 'merge' command can be used to generate | |
2251 an updated .isolate file out of it or the 'read' command to print it out to | |
2252 stdout. | |
2253 | |
2254 Argument processing stops at -- and these arguments are appended to the | |
2255 command line of the target to run. For example, use: | |
2256 isolate.py trace --isolated foo.isolated -- --gtest_filter=Foo.Bar | |
2257 """ | |
2258 add_trace_option(parser) | |
2259 parser.add_option( | |
2260 '-m', '--merge', action='store_true', | |
2261 help='After tracing, merge the results back in the .isolate file') | |
2262 parser.add_option( | |
2263 '--skip-refresh', action='store_true', | |
2264 help='Skip reading .isolate file and do not refresh the hash of ' | |
2265 'dependencies') | |
2266 options, args = parser.parse_args(args) | |
2267 | |
2268 complete_state = load_complete_state( | |
2269 options, os.getcwd(), None, options.skip_refresh) | |
2270 cmd = complete_state.saved_state.command + args | |
2271 if not cmd: | |
2272 raise ExecutionError('No command to run.') | |
2273 cmd = tools.fix_python_path(cmd) | |
2274 cwd = os.path.normpath(os.path.join( | |
2275 unicode(complete_state.root_dir), | |
2276 complete_state.saved_state.relative_cwd)) | |
2277 cmd[0] = os.path.normpath(os.path.join(cwd, cmd[0])) | |
2278 if not os.path.isfile(cmd[0]): | |
2279 raise ExecutionError( | |
2280 'Tracing failed for: %s\nIt doesn\'t exit' % ' '.join(cmd)) | |
2281 logging.info('Running %s, cwd=%s' % (cmd, cwd)) | |
2282 api = trace_inputs.get_api() | |
2283 logfile = complete_state.isolated_filepath + '.log' | |
2284 api.clean_trace(logfile) | |
2285 out = None | |
2286 try: | |
2287 with api.get_tracer(logfile) as tracer: | |
2288 result, out = tracer.trace( | |
2289 cmd, | |
2290 cwd, | |
2291 'default', | |
2292 True) | |
2293 except trace_inputs.TracingFailure, e: | |
2294 raise ExecutionError('Tracing failed for: %s\n%s' % (' '.join(cmd), str(e))) | |
2295 | |
2296 if result: | |
2297 logging.error( | |
2298 'Tracer exited with %d, which means the tests probably failed so the ' | |
2299 'trace is probably incomplete.', result) | |
2300 logging.info(out) | |
2301 | |
2302 complete_state.save_files() | |
2303 | |
2304 if options.merge: | |
2305 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist) | |
2306 merge(complete_state, blacklist) | |
2307 | |
2308 return result | |
2309 | |
2310 | |
2311 def _process_variable_arg(_option, _opt, _value, parser): | |
2312 if not parser.rargs: | |
2313 raise optparse.OptionValueError( | |
2314 'Please use --variable FOO=BAR or --variable FOO BAR') | |
2315 k = parser.rargs.pop(0) | |
2316 if '=' in k: | |
2317 parser.values.variables.append(tuple(k.split('=', 1))) | |
2318 else: | |
2319 if not parser.rargs: | |
2320 raise optparse.OptionValueError( | |
2321 'Please use --variable FOO=BAR or --variable FOO BAR') | |
2322 v = parser.rargs.pop(0) | |
2323 parser.values.variables.append((k, v)) | |
2324 | |
2325 | |
2326 def add_variable_option(parser): | |
2327 """Adds --isolated and --variable to an OptionParser.""" | |
2328 parser.add_option( | |
2329 '-s', '--isolated', | |
2330 metavar='FILE', | |
2331 help='.isolated file to generate or read') | |
2332 # Keep for compatibility. TODO(maruel): Remove once not used anymore. | |
2333 parser.add_option( | |
2334 '-r', '--result', | |
2335 dest='isolated', | |
2336 help=optparse.SUPPRESS_HELP) | |
2337 default_variables = [('OS', get_flavor())] | |
2338 if sys.platform in ('win32', 'cygwin'): | |
2339 default_variables.append(('EXECUTABLE_SUFFIX', '.exe')) | |
2340 else: | |
2341 default_variables.append(('EXECUTABLE_SUFFIX', '')) | |
2342 parser.add_option( | |
2343 '-V', '--variable', | |
2344 action='callback', | |
2345 callback=_process_variable_arg, | |
2346 default=default_variables, | |
2347 dest='variables', | |
2348 metavar='FOO BAR', | |
2349 help='Variables to process in the .isolate file, default: %default. ' | |
2350 'Variables are persistent accross calls, they are saved inside ' | |
2351 '<.isolated>.state') | |
2352 | |
2353 | |
2354 def add_trace_option(parser): | |
2355 """Adds --trace-blacklist to the parser.""" | |
2356 parser.add_option( | |
2357 '--trace-blacklist', | |
2358 action='append', default=list(DEFAULT_BLACKLIST), | |
2359 help='List of regexp to use as blacklist filter for files to consider ' | |
2360 'important, not to be confused with --blacklist which blacklists ' | |
2361 'test case.') | |
2362 | |
2363 | |
2364 def parse_isolated_option(parser, options, cwd, require_isolated): | |
2365 """Processes --isolated.""" | |
2366 if options.isolated: | |
2367 options.isolated = os.path.normpath( | |
2368 os.path.join(cwd, options.isolated.replace('/', os.path.sep))) | |
2369 if require_isolated and not options.isolated: | |
2370 parser.error('--isolated is required.') | |
2371 if options.isolated and not options.isolated.endswith('.isolated'): | |
2372 parser.error('--isolated value must end with \'.isolated\'') | |
2373 | |
2374 | |
2375 def parse_variable_option(options): | |
2376 """Processes --variable.""" | |
2377 # TODO(benrg): Maybe we should use a copy of gyp's NameValueListToDict here, | |
2378 # but it wouldn't be backward compatible. | |
2379 def try_make_int(s): | |
2380 """Converts a value to int if possible, converts to unicode otherwise.""" | |
2381 try: | |
2382 return int(s) | |
2383 except ValueError: | |
2384 return s.decode('utf-8') | |
2385 options.variables = dict((k, try_make_int(v)) for k, v in options.variables) | |
2386 | |
2387 | |
2388 class OptionParserIsolate(tools.OptionParserWithLogging): | |
2389 """Adds automatic --isolate, --isolated, --out and --variable handling.""" | |
2390 # Set it to False if it is not required, e.g. it can be passed on but do not | |
2391 # fail if not given. | |
2392 require_isolated = True | |
2393 | |
2394 def __init__(self, **kwargs): | |
2395 tools.OptionParserWithLogging.__init__( | |
2396 self, | |
2397 verbose=int(os.environ.get('ISOLATE_DEBUG', 0)), | |
2398 **kwargs) | |
2399 group = optparse.OptionGroup(self, "Common options") | |
2400 group.add_option( | |
2401 '-i', '--isolate', | |
2402 metavar='FILE', | |
2403 help='.isolate file to load the dependency data from') | |
2404 add_variable_option(group) | |
2405 group.add_option( | |
2406 '-o', '--outdir', metavar='DIR', | |
2407 help='Directory used to recreate the tree or store the hash table. ' | |
2408 'Defaults: run|remap: a /tmp subdirectory, others: ' | |
2409 'defaults to the directory containing --isolated') | |
2410 group.add_option( | |
2411 '--ignore_broken_items', action='store_true', | |
2412 default=bool(os.environ.get('ISOLATE_IGNORE_BROKEN_ITEMS')), | |
2413 help='Indicates that invalid entries in the isolated file to be ' | |
2414 'only be logged and not stop processing. Defaults to True if ' | |
2415 'env var ISOLATE_IGNORE_BROKEN_ITEMS is set') | |
2416 self.add_option_group(group) | |
2417 | |
2418 def parse_args(self, *args, **kwargs): | |
2419 """Makes sure the paths make sense. | |
2420 | |
2421 On Windows, / and \ are often mixed together in a path. | |
2422 """ | |
2423 options, args = tools.OptionParserWithLogging.parse_args( | |
2424 self, *args, **kwargs) | |
2425 if not self.allow_interspersed_args and args: | |
2426 self.error('Unsupported argument: %s' % args) | |
2427 | |
2428 cwd = file_path.get_native_path_case(unicode(os.getcwd())) | |
2429 parse_isolated_option(self, options, cwd, self.require_isolated) | |
2430 parse_variable_option(options) | |
2431 | |
2432 if options.isolate: | |
2433 # TODO(maruel): Work with non-ASCII. | |
2434 # The path must be in native path case for tracing purposes. | |
2435 options.isolate = unicode(options.isolate).replace('/', os.path.sep) | |
2436 options.isolate = os.path.normpath(os.path.join(cwd, options.isolate)) | |
2437 options.isolate = file_path.get_native_path_case(options.isolate) | |
2438 | |
2439 if options.outdir and not is_url(options.outdir): | |
2440 options.outdir = unicode(options.outdir).replace('/', os.path.sep) | |
2441 # outdir doesn't need native path case since tracing is never done from | |
2442 # there. | |
2443 options.outdir = os.path.normpath(os.path.join(cwd, options.outdir)) | |
2444 | |
2445 return options, args | |
2446 | |
2447 | |
2448 def main(argv): | |
2449 dispatcher = subcommand.CommandDispatcher(__name__) | |
2450 try: | |
2451 return dispatcher.execute(OptionParserIsolate(version=__version__), argv) | |
2452 except Exception as e: | |
2453 tools.report_error(e) | |
2454 return 1 | |
2455 | |
2456 | |
2457 if __name__ == '__main__': | |
2458 fix_encoding.fix_encoding() | |
2459 tools.disable_buffering() | |
2460 colorama.init() | |
2461 sys.exit(main(sys.argv[1:])) | |
OLD | NEW |