2 # Copyright (c) 2012 The Chromium Authors. All rights reserved.
3 # Use of this source code is governed by a BSD-style license that can be
4 # found in the LICENSE file.
6 """Front end tool to operate on .isolate files.
8 This includes creating, merging or compiling them to generate a .isolated file.
10 See more information at
11 https://code.google.com/p/swarming/wiki/IsolateDesign
12 https://code.google.com/p/swarming/wiki/IsolateUserGuide
14 # Run ./isolate.py --help for more detailed information.
32 # Import here directly so isolate is easier to use as a library.
33 from run_isolated import get_flavor
35 from third_party import colorama
36 from third_party.depot_tools import fix_encoding
37 from third_party.depot_tools import subcommand
39 from utils import file_path
40 from utils import tools
41 from utils import short_expression_finder
47 PATH_VARIABLES = ('DEPTH', 'PRODUCT_DIR')
49 # Files that should be 0-length when mapped.
50 KEY_TOUCHED = 'isolate_dependency_touched'
51 # Files that should be tracked by the build tool.
52 KEY_TRACKED = 'isolate_dependency_tracked'
53 # Files that should not be tracked by the build tool.
54 KEY_UNTRACKED = 'isolate_dependency_untracked'
57 class ExecutionError(Exception):
58 """A generic error occurred."""
63 ### Path handling code.
67 # Temporary vim or python files.
69 # .git or .svn directory.
70 r'^(?:.+' + re.escape(os.path.sep) + r'|)\.(?:git|svn)$',
75 DEFAULT_BLACKLIST += (
76 r'^.+\.(?:run_test_cases)$',
77 r'^(?:.+' + re.escape(os.path.sep) + r'|)testserver\.log$',
81 def relpath(path, root):
82 """os.path.relpath() that keeps trailing os.path.sep."""
83 out = os.path.relpath(path, root)
84 if path.endswith(os.path.sep):
89 def safe_relpath(filepath, basepath):
90 """Do not throw on Windows when filepath and basepath are on different drives.
92 Different than relpath() above since this one doesn't keep the trailing
93 os.path.sep and it swallows exceptions on Windows and return the original
94 absolute path in the case of different drives.
97 return os.path.relpath(filepath, basepath)
99 assert sys.platform == 'win32'
104 """os.path.normpath() that keeps trailing os.path.sep."""
105 out = os.path.normpath(path)
106 if path.endswith(os.path.sep):
111 def posix_relpath(path, root):
112 """posix.relpath() that keeps trailing slash."""
113 out = posixpath.relpath(path, root)
114 if path.endswith('/'):
120 """Cleans up a relative path. Converts any os.path.sep to '/' on Windows."""
122 x = x.rstrip(os.path.sep).replace(os.path.sep, '/')
131 return bool(re.match(r'^https?://.+$', path))
134 def path_starts_with(prefix, path):
135 """Returns true if the components of the path |prefix| are the same as the
136 initial components of |path| (or all of the components of |path|). The paths
139 assert os.path.isabs(prefix) and os.path.isabs(path)
140 prefix = os.path.normpath(prefix)
141 path = os.path.normpath(path)
142 assert prefix == file_path.get_native_path_case(prefix), prefix
143 assert path == file_path.get_native_path_case(path), path
144 prefix = prefix.rstrip(os.path.sep) + os.path.sep
145 path = path.rstrip(os.path.sep) + os.path.sep
146 return path.startswith(prefix)
149 def fix_native_path_case(root, path):
150 """Ensures that each component of |path| has the proper native case by
151 iterating slowly over the directory elements of |path|."""
152 native_case_path = root
153 for raw_part in path.split(os.sep):
154 if not raw_part or raw_part == '.':
157 part = file_path.find_item_native_case(native_case_path, raw_part)
159 raise isolateserver.MappingError(
160 'Input file %s doesn\'t exist' %
161 os.path.join(native_case_path, raw_part))
162 native_case_path = os.path.join(native_case_path, part)
164 return os.path.normpath(native_case_path)
167 def expand_symlinks(indir, relfile):
168 """Follows symlinks in |relfile|, but treating symlinks that point outside the
169 build tree as if they were ordinary directories/files. Returns the final
170 symlink-free target and a list of paths to symlinks encountered in the
173 The rule about symlinks outside the build tree is for the benefit of the
174 Chromium OS ebuild, which symlinks the output directory to an unrelated path
177 Fails when a directory loop is detected, although in theory we could support
180 is_directory = relfile.endswith(os.path.sep)
182 todo = relfile.strip(os.path.sep)
186 pre_symlink, symlink, post_symlink = file_path.split_at_symlink(
189 todo = fix_native_path_case(done, todo)
190 done = os.path.join(done, todo)
192 symlink_path = os.path.join(done, pre_symlink, symlink)
193 post_symlink = post_symlink.lstrip(os.path.sep)
194 # readlink doesn't exist on Windows.
195 # pylint: disable=E1101
196 target = os.path.normpath(os.path.join(done, pre_symlink))
197 symlink_target = os.readlink(symlink_path)
198 if os.path.isabs(symlink_target):
199 # Absolute path are considered a normal directories. The use case is
200 # generally someone who puts the output directory on a separate drive.
201 target = symlink_target
203 # The symlink itself could be using the wrong path case.
204 target = fix_native_path_case(target, symlink_target)
206 if not os.path.exists(target):
207 raise isolateserver.MappingError(
208 'Symlink target doesn\'t exist: %s -> %s' % (symlink_path, target))
209 target = file_path.get_native_path_case(target)
210 if not path_starts_with(indir, target):
214 if path_starts_with(target, symlink_path):
215 raise isolateserver.MappingError(
216 'Can\'t map recursive symlink reference %s -> %s' %
217 (symlink_path, target))
218 logging.info('Found symlink: %s -> %s', symlink_path, target)
219 symlinks.append(os.path.relpath(symlink_path, indir))
220 # Treat the common prefix of the old and new paths as done, and start
222 target = target.split(os.path.sep)
223 symlink_path = symlink_path.split(os.path.sep)
225 for target_piece, symlink_path_piece in zip(target, symlink_path):
226 if target_piece == symlink_path_piece:
230 done = os.path.sep.join(target[:prefix_length])
232 os.path.sep.join(target[prefix_length:]), post_symlink)
234 relfile = os.path.relpath(done, indir)
235 relfile = relfile.rstrip(os.path.sep) + is_directory * os.path.sep
236 return relfile, symlinks
239 def expand_directory_and_symlink(indir, relfile, blacklist, follow_symlinks):
240 """Expands a single input. It can result in multiple outputs.
242 This function is recursive when relfile is a directory.
244 Note: this code doesn't properly handle recursive symlink like one created
248 if os.path.isabs(relfile):
249 raise isolateserver.MappingError(
250 'Can\'t map absolute path %s' % relfile)
252 infile = normpath(os.path.join(indir, relfile))
253 if not infile.startswith(indir):
254 raise isolateserver.MappingError(
255 'Can\'t map file %s outside %s' % (infile, indir))
257 filepath = os.path.join(indir, relfile)
258 native_filepath = file_path.get_native_path_case(filepath)
259 if filepath != native_filepath:
261 if filepath != native_filepath + '.' + os.path.sep:
262 # Give up enforcing strict path case on OSX. Really, it's that sad. The
263 # case where it happens is very specific and hard to reproduce:
264 # get_native_path_case(
265 # u'Foo.framework/Versions/A/Resources/Something.nib') will return
266 # u'Foo.framework/Versions/A/resources/Something.nib', e.g. lowercase 'r'.
268 # Note that this is really something deep in OSX because running
269 # ls Foo.framework/Versions/A
270 # will print out 'Resources', while file_path.get_native_path_case()
271 # returns a lower case 'r'.
273 # So *something* is happening under the hood resulting in the command 'ls'
274 # and Carbon.File.FSPathMakeRef('path').FSRefMakePath() to disagree. We
276 if sys.platform != 'darwin':
277 raise isolateserver.MappingError(
278 'File path doesn\'t equal native file path\n%s != %s' %
279 (filepath, native_filepath))
283 relfile, symlinks = expand_symlinks(indir, relfile)
285 if relfile.endswith(os.path.sep):
286 if not os.path.isdir(infile):
287 raise isolateserver.MappingError(
288 '%s is not a directory but ends with "%s"' % (infile, os.path.sep))
291 if relfile.startswith('.' + os.path.sep):
292 relfile = relfile[2:]
295 for filename in os.listdir(infile):
296 inner_relfile = os.path.join(relfile, filename)
297 if blacklist(inner_relfile):
299 if os.path.isdir(os.path.join(indir, inner_relfile)):
300 inner_relfile += os.path.sep
302 expand_directory_and_symlink(indir, inner_relfile, blacklist,
306 raise isolateserver.MappingError(
307 'Unable to iterate over directory %s.\n%s' % (infile, e))
309 # Always add individual files even if they were blacklisted.
310 if os.path.isdir(infile):
311 raise isolateserver.MappingError(
312 'Input directory %s must have a trailing slash' % infile)
314 if not os.path.isfile(infile):
315 raise isolateserver.MappingError(
316 'Input file %s doesn\'t exist' % infile)
318 return symlinks + [relfile]
321 def expand_directories_and_symlinks(indir, infiles, blacklist,
322 follow_symlinks, ignore_broken_items):
323 """Expands the directories and the symlinks, applies the blacklist and
324 verifies files exist.
326 Files are specified in os native path separator.
329 for relfile in infiles:
331 outfiles.extend(expand_directory_and_symlink(indir, relfile, blacklist,
333 except isolateserver.MappingError as e:
334 if ignore_broken_items:
335 logging.info('warning: %s', e)
341 def recreate_tree(outdir, indir, infiles, action, as_hash):
342 """Creates a new tree with only the input files in it.
345 outdir: Output directory to create the files in.
346 indir: Root directory the infiles are based in.
347 infiles: dict of files to map from |indir| to |outdir|.
348 action: One of accepted action of run_isolated.link_file().
349 as_hash: Output filename is the hash instead of relfile.
352 'recreate_tree(outdir=%s, indir=%s, files=%d, action=%s, as_hash=%s)' %
353 (outdir, indir, len(infiles), action, as_hash))
355 assert os.path.isabs(outdir) and outdir == os.path.normpath(outdir), outdir
356 if not os.path.isdir(outdir):
357 logging.info('Creating %s' % outdir)
360 for relfile, metadata in infiles.iteritems():
361 infile = os.path.join(indir, relfile)
363 # Do the hashtable specific checks.
365 # Skip links when storing a hashtable.
367 outfile = os.path.join(outdir, metadata['h'])
368 if os.path.isfile(outfile):
369 # Just do a quick check that the file size matches. No need to stat()
370 # again the input file, grab the value from the dict.
371 if not 's' in metadata:
372 raise isolateserver.MappingError(
373 'Misconfigured item %s: %s' % (relfile, metadata))
374 if metadata['s'] == os.stat(outfile).st_size:
377 logging.warn('Overwritting %s' % metadata['h'])
380 outfile = os.path.join(outdir, relfile)
381 outsubdir = os.path.dirname(outfile)
382 if not os.path.isdir(outsubdir):
383 os.makedirs(outsubdir)
385 # TODO(csharp): Fix crbug.com/150823 and enable the touched logic again.
386 # if metadata.get('T') == True:
387 # open(outfile, 'ab').close()
389 pointed = metadata['l']
390 logging.debug('Symlink: %s -> %s' % (outfile, pointed))
391 # symlink doesn't exist on Windows.
392 os.symlink(pointed, outfile) # pylint: disable=E1101
394 run_isolated.link_file(outfile, infile, action)
397 def process_input(filepath, prevdict, read_only, flavor, algo):
398 """Processes an input file, a dependency, and return meta data about it.
401 - filepath: File to act on.
402 - prevdict: the previous dictionary. It is used to retrieve the cached sha-1
403 to skip recalculating the hash.
404 - read_only: If True, the file mode is manipulated. In practice, only save
405 one of 4 modes: 0755 (rwx), 0644 (rw), 0555 (rx), 0444 (r). On
406 windows, mode is not set since all files are 'executable' by
408 - algo: Hashing algorithm used.
411 - Retrieves the file mode, file size, file timestamp, file link
412 destination if it is a file link and calcultate the SHA-1 of the file's
413 content if the path points to a file and not a symlink.
416 # TODO(csharp): Fix crbug.com/150823 and enable the touched logic again.
417 # if prevdict.get('T') == True:
418 # # The file's content is ignored. Skip the time and hard code mode.
419 # if get_flavor() != 'win':
420 # out['m'] = stat.S_IRUSR | stat.S_IRGRP
422 # out['h'] = algo().hexdigest()
426 # Always check the file stat and check if it is a link. The timestamp is used
427 # to know if the file's content/symlink destination should be looked into.
428 # E.g. only reuse from prevdict if the timestamp hasn't changed.
429 # There is the risk of the file's timestamp being reset to its last value
430 # manually while its content changed. We don't protect against that use case.
432 filestats = os.lstat(filepath)
434 # The file is not present.
435 raise isolateserver.MappingError('%s is missing' % filepath)
436 is_link = stat.S_ISLNK(filestats.st_mode)
439 # Ignore file mode on Windows since it's not really useful there.
440 filemode = stat.S_IMODE(filestats.st_mode)
441 # Remove write access for group and all access to 'others'.
442 filemode &= ~(stat.S_IWGRP | stat.S_IRWXO)
444 filemode &= ~stat.S_IWUSR
445 if filemode & stat.S_IXUSR:
446 filemode |= stat.S_IXGRP
448 filemode &= ~stat.S_IXGRP
452 # Used to skip recalculating the hash or link destination. Use the most recent
454 # TODO(maruel): Save it in the .state file instead of .isolated so the
455 # .isolated file is deterministic.
456 out['t'] = int(round(filestats.st_mtime))
459 out['s'] = filestats.st_size
460 # If the timestamp wasn't updated and the file size is still the same, carry
462 if (prevdict.get('t') == out['t'] and
463 prevdict.get('s') == out['s']):
464 # Reuse the previous hash if available.
465 out['h'] = prevdict.get('h')
467 out['h'] = isolateserver.hash_file(filepath, algo)
469 # If the timestamp wasn't updated, carry on the link destination.
470 if prevdict.get('t') == out['t']:
471 # Reuse the previous link destination if available.
472 out['l'] = prevdict.get('l')
473 if out.get('l') is None:
474 # The link could be in an incorrect path case. In practice, this only
475 # happen on OSX on case insensitive HFS.
476 # TODO(maruel): It'd be better if it was only done once, in
477 # expand_directory_and_symlink(), so it would not be necessary to do again
479 symlink_value = os.readlink(filepath) # pylint: disable=E1101
480 filedir = file_path.get_native_path_case(os.path.dirname(filepath))
481 native_dest = fix_native_path_case(filedir, symlink_value)
482 out['l'] = os.path.relpath(native_dest, filedir)
489 def isolatedfile_to_state(filename):
490 """Replaces the file's extension."""
491 return filename + '.state'
494 def determine_root_dir(relative_root, infiles):
495 """For a list of infiles, determines the deepest root directory that is
496 referenced indirectly.
498 All arguments must be using os.path.sep.
500 # The trick used to determine the root directory is to look at "how far" back
501 # up it is looking up.
502 deepest_root = relative_root
505 while i.startswith('..' + os.path.sep):
507 assert not i.startswith(os.path.sep)
508 x = os.path.dirname(x)
509 if deepest_root.startswith(x):
512 'determine_root_dir(%s, %d files) -> %s' % (
513 relative_root, len(infiles), deepest_root))
517 def replace_variable(part, variables):
518 m = re.match(r'<\(([A-Z_]+)\)', part)
520 if m.group(1) not in variables:
521 raise ExecutionError(
522 'Variable "%s" was not found in %s.\nDid you forget to specify '
523 '--variable?' % (m.group(1), variables))
524 return variables[m.group(1)]
528 def process_variables(cwd, variables, relative_base_dir):
529 """Processes path variables as a special case and returns a copy of the dict.
531 For each 'path' variable: first normalizes it based on |cwd|, verifies it
532 exists then sets it as relative to relative_base_dir.
534 relative_base_dir = file_path.get_native_path_case(relative_base_dir)
535 variables = variables.copy()
536 for i in PATH_VARIABLES:
537 if i not in variables:
539 variable = variables[i].strip()
540 # Variables could contain / or \ on windows. Always normalize to
542 variable = variable.replace('/', os.path.sep)
543 variable = os.path.join(cwd, variable)
544 variable = os.path.normpath(variable)
545 variable = file_path.get_native_path_case(variable)
546 if not os.path.isdir(variable):
547 raise ExecutionError('%s=%s is not a directory' % (i, variable))
549 # All variables are relative to the .isolate file.
550 variable = os.path.relpath(variable, relative_base_dir)
552 'Translated variable %s from %s to %s', i, variables[i], variable)
553 variables[i] = variable
557 def eval_variables(item, variables):
558 """Replaces the .isolate variables in a string item.
560 Note that the .isolate format is a subset of the .gyp dialect.
563 replace_variable(p, variables) for p in re.split(r'(<\([A-Z_]+\))', item))
566 def classify_files(root_dir, tracked, untracked):
567 """Converts the list of files into a .isolate 'variables' dictionary.
570 - tracked: list of files names to generate a dictionary out of that should
572 - untracked: list of files names that must not be tracked.
574 # These directories are not guaranteed to be always present on every builder.
575 OPTIONAL_DIRECTORIES = (
577 'third_party/WebKit/LayoutTests',
581 new_untracked = list(untracked)
583 def should_be_tracked(filepath):
584 """Returns True if it is a file without whitespace in a non-optional
585 directory that has no symlink in its path.
587 if filepath.endswith('/'):
591 if any(i in filepath for i in OPTIONAL_DIRECTORIES):
593 # Look if any element in the path is a symlink.
594 split = filepath.split('/')
595 for i in range(len(split)):
596 if os.path.islink(os.path.join(root_dir, '/'.join(split[:i+1]))):
600 for filepath in sorted(tracked):
601 if should_be_tracked(filepath):
602 new_tracked.append(filepath)
605 new_untracked.append(filepath)
609 variables[KEY_TRACKED] = sorted(new_tracked)
611 variables[KEY_UNTRACKED] = sorted(new_untracked)
615 def chromium_fix(f, variables):
616 """Fixes an isolate dependnecy with Chromium-specific fixes."""
617 # Skip log in PRODUCT_DIR. Note that these are applied on '/' style path
619 LOG_FILE = re.compile(r'^\<\(PRODUCT_DIR\)\/[^\/]+\.log$')
622 # http://crbug.com/160539, on Windows, it's in chrome/.
624 'chrome/Media Cache/',
625 # 'First Run' is not created by the compile, but by the test itself.
626 '<(PRODUCT_DIR)/First Run')
628 # Blacklist logs and other unimportant files.
629 if LOG_FILE.match(f) or f in IGNORED_ITEMS:
630 logging.debug('Ignoring %s', f)
633 EXECUTABLE = re.compile(
634 r'^(\<\(PRODUCT_DIR\)\/[^\/\.]+)' +
635 re.escape(variables.get('EXECUTABLE_SUFFIX', '')) +
637 match = EXECUTABLE.match(f)
639 return match.group(1) + '<(EXECUTABLE_SUFFIX)'
641 if sys.platform == 'darwin':
642 # On OSX, the name of the output is dependent on gyp define, it can be
643 # 'Google Chrome.app' or 'Chromium.app', same for 'XXX
644 # Framework.framework'. Furthermore, they are versioned with a gyp
645 # variable. To lower the complexity of the .isolate file, remove all the
646 # individual entries that show up under any of the 4 entries and replace
647 # them with the directory itself. Overall, this results in a bit more
648 # files than strictly necessary.
650 '<(PRODUCT_DIR)/Chromium Framework.framework/',
651 '<(PRODUCT_DIR)/Chromium.app/',
652 '<(PRODUCT_DIR)/Google Chrome Framework.framework/',
653 '<(PRODUCT_DIR)/Google Chrome.app/',
655 for prefix in OSX_BUNDLES:
656 if f.startswith(prefix):
657 # Note this result in duplicate values, so the a set() must be used to
663 def generate_simplified(
664 tracked, untracked, touched, root_dir, variables, relative_cwd,
666 """Generates a clean and complete .isolate 'variables' dictionary.
668 Cleans up and extracts only files from within root_dir then processes
669 variables and relative_cwd.
671 root_dir = os.path.realpath(root_dir)
673 'generate_simplified(%d files, %s, %s, %s)' %
674 (len(tracked) + len(untracked) + len(touched),
675 root_dir, variables, relative_cwd))
678 relative_cwd = cleanup_path(relative_cwd)
679 assert not os.path.isabs(relative_cwd), relative_cwd
680 # Creates the right set of variables here. We only care about PATH_VARIABLES.
681 path_variables = dict(
682 ('<(%s)' % k, variables[k].replace(os.path.sep, '/'))
683 for k in PATH_VARIABLES if k in variables)
684 variables = variables.copy()
685 variables.update(path_variables)
687 # Actual work: Process the files.
688 # TODO(maruel): if all the files in a directory are in part tracked and in
689 # part untracked, the directory will not be extracted. Tracked files should be
690 # 'promoted' to be untracked as needed.
691 tracked = trace_inputs.extract_directories(
692 root_dir, tracked, trace_blacklist)
693 untracked = trace_inputs.extract_directories(
694 root_dir, untracked, trace_blacklist)
695 # touched is not compressed, otherwise it would result in files to be archived
696 # that we don't need.
698 root_dir_posix = root_dir.replace(os.path.sep, '/')
700 """Bases the file on the most restrictive variable."""
701 # Important, GYP stores the files with / and not \.
702 f = f.replace(os.path.sep, '/')
703 logging.debug('fix(%s)' % f)
704 # If it's not already a variable.
705 if not f.startswith('<'):
706 # relative_cwd is usually the directory containing the gyp file. It may be
707 # empty if the whole directory containing the gyp file is needed.
708 # Use absolute paths in case cwd_dir is outside of root_dir.
709 # Convert the whole thing to / since it's isolate's speak.
711 posixpath.join(root_dir_posix, f),
712 posixpath.join(root_dir_posix, relative_cwd)) or './'
714 for variable, root_path in path_variables.iteritems():
715 if f.startswith(root_path):
716 f = variable + f[len(root_path):]
717 logging.debug('Converted to %s' % f)
722 """Reduces the items to convert variables, removes unneeded items, apply
723 chromium-specific fixes and only return unique items.
725 variables_converted = (fix(f.path) for f in items)
726 chromium_fixed = (chromium_fix(f, variables) for f in variables_converted)
727 return set(f for f in chromium_fixed if f)
729 tracked = fix_all(tracked)
730 untracked = fix_all(untracked)
731 touched = fix_all(touched)
732 out = classify_files(root_dir, tracked, untracked)
734 out[KEY_TOUCHED] = sorted(touched)
738 def chromium_filter_flags(variables):
739 """Filters out build flags used in Chromium that we don't want to treat as
740 configuration variables.
742 # TODO(benrg): Need a better way to determine this.
743 blacklist = set(PATH_VARIABLES + ('EXECUTABLE_SUFFIX', 'FLAG'))
744 return dict((k, v) for k, v in variables.iteritems() if k not in blacklist)
747 def generate_isolate(
748 tracked, untracked, touched, root_dir, variables, relative_cwd,
750 """Generates a clean and complete .isolate file."""
751 dependencies = generate_simplified(
752 tracked, untracked, touched, root_dir, variables, relative_cwd,
754 config_variables = chromium_filter_flags(variables)
755 config_variable_names, config_values = zip(
756 *sorted(config_variables.iteritems()))
758 # The new dependencies apply to just one configuration, namely config_values.
759 out.merge_dependencies(dependencies, config_variable_names, [config_values])
760 return out.make_isolate_file()
763 def split_touched(files):
764 """Splits files that are touched vs files that are read."""
772 return tracked, touched
775 def pretty_print(variables, stdout):
776 """Outputs a gyp compatible list from the decoded variables.
778 Similar to pprint.print() but with NIH syndrome.
780 # Order the dictionary keys by these keys in priority.
782 'variables', 'condition', 'command', 'relative_cwd', 'read_only',
783 KEY_TRACKED, KEY_UNTRACKED)
786 """Gives priority to 'most important' keys before the others."""
788 return str(ORDER.index(x))
791 def loop_list(indent, items):
793 if isinstance(item, basestring):
794 stdout.write('%s\'%s\',\n' % (indent, item))
795 elif isinstance(item, dict):
796 stdout.write('%s{\n' % indent)
797 loop_dict(indent + ' ', item)
798 stdout.write('%s},\n' % indent)
799 elif isinstance(item, list):
800 # A list inside a list will write the first item embedded.
801 stdout.write('%s[' % indent)
802 for index, i in enumerate(item):
803 if isinstance(i, basestring):
805 '\'%s\', ' % i.replace('\\', '\\\\').replace('\'', '\\\''))
806 elif isinstance(i, dict):
808 loop_dict(indent + ' ', i)
809 if index != len(item) - 1:
813 stdout.write('%s}%s' % (indent, x))
820 def loop_dict(indent, items):
821 for key in sorted(items, key=sorting_key):
823 stdout.write("%s'%s': " % (indent, key))
824 if isinstance(item, dict):
826 loop_dict(indent + ' ', item)
827 stdout.write(indent + '},\n')
828 elif isinstance(item, list):
830 loop_list(indent + ' ', item)
831 stdout.write(indent + '],\n')
832 elif isinstance(item, basestring):
834 '\'%s\',\n' % item.replace('\\', '\\\\').replace('\'', '\\\''))
835 elif item in (True, False, None):
836 stdout.write('%s\n' % item)
841 loop_dict(' ', variables)
846 """Merges two compatible datastructures composed of dict/list/set."""
847 assert lhs is not None or rhs is not None
849 return copy.deepcopy(rhs)
851 return copy.deepcopy(lhs)
852 assert type(lhs) == type(rhs), (lhs, rhs)
853 if hasattr(lhs, 'union'):
854 # Includes set, ConfigSettings and Configs.
855 return lhs.union(rhs)
856 if isinstance(lhs, dict):
857 return dict((k, union(lhs.get(k), rhs.get(k))) for k in set(lhs).union(rhs))
858 elif isinstance(lhs, list):
859 # Do not go inside the list.
861 assert False, type(lhs)
864 def extract_comment(content):
865 """Extracts file level comment."""
867 for line in content.splitlines(True):
868 if line.startswith('#'):
875 def eval_content(content):
876 """Evaluates a python file and return the value defined in it.
878 Used in practice for .isolate files.
880 globs = {'__builtins__': None}
883 value = eval(content, globs, locs)
884 except TypeError as e:
885 e.args = list(e.args) + [content]
887 assert locs == {}, locs
888 assert globs == {'__builtins__': None}, globs
892 def match_configs(expr, config_variables, all_configs):
893 """Returns the configs from |all_configs| that match the |expr|, where
894 the elements of |all_configs| are tuples of values for the |config_variables|.
896 >>> match_configs(expr = "(foo==1 or foo==2) and bar=='b'",
897 config_variables = ["foo", "bar"],
898 all_configs = [(1, 'a'), (1, 'b'), (2, 'a'), (2, 'b')])
902 config for config in all_configs
903 if eval(expr, dict(zip(config_variables, config)))
907 def verify_variables(variables):
908 """Verifies the |variables| dictionary is in the expected format."""
916 assert isinstance(variables, dict), variables
917 assert set(VALID_VARIABLES).issuperset(set(variables)), variables.keys()
918 for name, value in variables.iteritems():
919 if name == 'read_only':
920 assert value in (True, False, None), value
922 assert isinstance(value, list), value
923 assert all(isinstance(i, basestring) for i in value), value
926 def verify_ast(expr, variables_and_values):
927 """Verifies that |expr| is of the form
928 expr ::= expr ( "or" | "and" ) expr
929 | identifier "==" ( string | int )
930 Also collects the variable identifiers and string/int values in the dict
931 |variables_and_values|, in the form {'var': set([val1, val2, ...]), ...}.
933 assert isinstance(expr, (ast.BoolOp, ast.Compare))
934 if isinstance(expr, ast.BoolOp):
935 assert isinstance(expr.op, (ast.And, ast.Or))
936 for subexpr in expr.values:
937 verify_ast(subexpr, variables_and_values)
939 assert isinstance(expr.left.ctx, ast.Load)
940 assert len(expr.ops) == 1
941 assert isinstance(expr.ops[0], ast.Eq)
942 var_values = variables_and_values.setdefault(expr.left.id, set())
943 rhs = expr.comparators[0]
944 assert isinstance(rhs, (ast.Str, ast.Num))
945 var_values.add(rhs.n if isinstance(rhs, ast.Num) else rhs.s)
948 def verify_condition(condition, variables_and_values):
949 """Verifies the |condition| dictionary is in the expected format.
950 See verify_ast() for the meaning of |variables_and_values|.
952 VALID_INSIDE_CONDITION = ['variables']
953 assert isinstance(condition, list), condition
954 assert len(condition) == 2, condition
955 expr, then = condition
957 test_ast = compile(expr, '<condition>', 'eval', ast.PyCF_ONLY_AST)
958 verify_ast(test_ast.body, variables_and_values)
960 assert isinstance(then, dict), then
961 assert set(VALID_INSIDE_CONDITION).issuperset(set(then)), then.keys()
962 verify_variables(then['variables'])
965 def verify_root(value, variables_and_values):
966 """Verifies that |value| is the parsed form of a valid .isolate file.
967 See verify_ast() for the meaning of |variables_and_values|.
969 VALID_ROOTS = ['includes', 'conditions']
970 assert isinstance(value, dict), value
971 assert set(VALID_ROOTS).issuperset(set(value)), value.keys()
973 includes = value.get('includes', [])
974 assert isinstance(includes, list), includes
975 for include in includes:
976 assert isinstance(include, basestring), include
978 conditions = value.get('conditions', [])
979 assert isinstance(conditions, list), conditions
980 for condition in conditions:
981 verify_condition(condition, variables_and_values)
984 def remove_weak_dependencies(values, key, item, item_configs):
985 """Removes any configs from this key if the item is already under a
988 if key == KEY_TOUCHED:
989 item_configs = set(item_configs)
990 for stronger_key in (KEY_TRACKED, KEY_UNTRACKED):
992 item_configs -= values[stronger_key][item]
999 def remove_repeated_dependencies(folders, key, item, item_configs):
1000 """Removes any configs from this key if the item is in a folder that is
1001 already included."""
1003 if key in (KEY_UNTRACKED, KEY_TRACKED, KEY_TOUCHED):
1004 item_configs = set(item_configs)
1005 for (folder, configs) in folders.iteritems():
1006 if folder != item and item.startswith(folder):
1007 item_configs -= configs
1012 def get_folders(values_dict):
1013 """Returns a dict of all the folders in the given value_dict."""
1015 (item, configs) for (item, configs) in values_dict.iteritems()
1016 if item.endswith('/')
1020 def invert_map(variables):
1021 """Converts {config: {deptype: list(depvals)}} to
1022 {deptype: {depval: set(configs)}}.
1031 out = dict((key, {}) for key in KEYS)
1032 for config, values in variables.iteritems():
1034 if key == 'command':
1035 items = [tuple(values[key])] if key in values else []
1036 elif key == 'read_only':
1037 items = [values[key]] if key in values else []
1039 assert key in (KEY_TOUCHED, KEY_TRACKED, KEY_UNTRACKED)
1040 items = values.get(key, [])
1042 out[key].setdefault(item, set()).add(config)
1046 def reduce_inputs(values):
1047 """Reduces the output of invert_map() to the strictest minimum list.
1049 Looks at each individual file and directory, maps where they are used and
1050 reconstructs the inverse dictionary.
1052 Returns the minimized dictionary.
1062 # Folders can only live in KEY_UNTRACKED.
1063 folders = get_folders(values.get(KEY_UNTRACKED, {}))
1065 out = dict((key, {}) for key in KEYS)
1067 for item, item_configs in values.get(key, {}).iteritems():
1068 item_configs = remove_weak_dependencies(values, key, item, item_configs)
1069 item_configs = remove_repeated_dependencies(
1070 folders, key, item, item_configs)
1072 out[key][item] = item_configs
1076 def convert_map_to_isolate_dict(values, config_variables):
1077 """Regenerates back a .isolate configuration dict from files and dirs
1078 mappings generated from reduce_inputs().
1080 # Gather a list of configurations for set inversion later.
1081 all_mentioned_configs = set()
1082 for configs_by_item in values.itervalues():
1083 for configs in configs_by_item.itervalues():
1084 all_mentioned_configs.update(configs)
1086 # Invert the mapping to make it dict first.
1089 for item, configs in values[key].iteritems():
1090 then = conditions.setdefault(frozenset(configs), {})
1091 variables = then.setdefault('variables', {})
1093 if item in (True, False):
1094 # One-off for read_only.
1095 variables[key] = item
1098 if isinstance(item, tuple):
1099 # One-off for command.
1100 # Do not merge lists and do not sort!
1101 # Note that item is a tuple.
1102 assert key not in variables
1103 variables[key] = list(item)
1105 # The list of items (files or dirs). Append the new item and keep
1107 l = variables.setdefault(key, [])
1111 if all_mentioned_configs:
1112 config_values = map(set, zip(*all_mentioned_configs))
1113 sef = short_expression_finder.ShortExpressionFinder(
1114 zip(config_variables, config_values))
1116 conditions = sorted(
1117 [sef.get_expr(configs), then] for configs, then in conditions.iteritems())
1118 return {'conditions': conditions}
1121 ### Internal state files.
1124 class ConfigSettings(object):
1125 """Represents the dependency variables for a single build configuration.
1126 The structure is immutable.
1128 def __init__(self, config, values):
1129 self.config = config
1130 verify_variables(values)
1131 self.touched = sorted(values.get(KEY_TOUCHED, []))
1132 self.tracked = sorted(values.get(KEY_TRACKED, []))
1133 self.untracked = sorted(values.get(KEY_UNTRACKED, []))
1134 self.command = values.get('command', [])[:]
1135 self.read_only = values.get('read_only')
1137 def union(self, rhs):
1138 assert not (self.config and rhs.config) or (self.config == rhs.config)
1139 assert not (self.command and rhs.command) or (self.command == rhs.command)
1141 KEY_TOUCHED: sorted(self.touched + rhs.touched),
1142 KEY_TRACKED: sorted(self.tracked + rhs.tracked),
1143 KEY_UNTRACKED: sorted(self.untracked + rhs.untracked),
1144 'command': self.command or rhs.command,
1145 'read_only': rhs.read_only if self.read_only is None else self.read_only,
1147 return ConfigSettings(self.config or rhs.config, var)
1152 out['command'] = self.command
1154 out[KEY_TOUCHED] = self.touched
1156 out[KEY_TRACKED] = self.tracked
1158 out[KEY_UNTRACKED] = self.untracked
1159 if self.read_only is not None:
1160 out['read_only'] = self.read_only
1164 class Configs(object):
1165 """Represents a processed .isolate file.
1167 Stores the file in a processed way, split by configuration.
1169 def __init__(self, file_comment):
1170 self.file_comment = file_comment
1171 # The keys of by_config are tuples of values for the configuration
1172 # variables. The names of the variables (which must be the same for
1173 # every by_config key) are kept in config_variables. Initially by_config
1174 # is empty and we don't know what configuration variables will be used,
1175 # so config_variables also starts out empty. It will be set by the first
1176 # call to union() or merge_dependencies().
1178 self.config_variables = ()
1180 def union(self, rhs):
1181 """Adds variables from rhs (a Configs) to the existing variables.
1183 config_variables = self.config_variables
1184 if not config_variables:
1185 config_variables = rhs.config_variables
1187 # We can't proceed if this isn't true since we don't know the correct
1188 # default values for extra variables. The variables are sorted so we
1189 # don't need to worry about permutations.
1190 if rhs.config_variables and rhs.config_variables != config_variables:
1191 raise ExecutionError(
1192 'Variables in merged .isolate files do not match: %r and %r' % (
1193 config_variables, rhs.config_variables))
1195 # Takes the first file comment, prefering lhs.
1196 out = Configs(self.file_comment or rhs.file_comment)
1197 out.config_variables = config_variables
1198 for config in set(self.by_config) | set(rhs.by_config):
1199 out.by_config[config] = union(
1200 self.by_config.get(config), rhs.by_config.get(config))
1203 def merge_dependencies(self, values, config_variables, configs):
1204 """Adds new dependencies to this object for the given configurations.
1206 values: A variables dict as found in a .isolate file, e.g.,
1207 {KEY_TOUCHED: [...], 'command': ...}.
1208 config_variables: An ordered list of configuration variables, e.g.,
1209 ["OS", "chromeos"]. If this object already contains any dependencies,
1210 the configuration variables must match.
1211 configs: a list of tuples of values of the configuration variables,
1212 e.g., [("mac", 0), ("linux", 1)]. The dependencies in |values|
1213 are added to all of these configurations, and other configurations
1219 if not self.config_variables:
1220 self.config_variables = config_variables
1222 # See comment in Configs.union().
1223 assert self.config_variables == config_variables
1225 for config in configs:
1226 self.by_config[config] = union(
1227 self.by_config.get(config), ConfigSettings(config, values))
1230 """Returns a flat dictionary representation of the configuration.
1232 return dict((k, v.flatten()) for k, v in self.by_config.iteritems())
1234 def make_isolate_file(self):
1235 """Returns a dictionary suitable for writing to a .isolate file.
1237 dependencies_by_config = self.flatten()
1238 configs_by_dependency = reduce_inputs(invert_map(dependencies_by_config))
1239 return convert_map_to_isolate_dict(configs_by_dependency,
1240 self.config_variables)
1243 # TODO(benrg): Remove this function when no old-format files are left.
1244 def convert_old_to_new_format(value):
1245 """Converts from the old .isolate format, which only has one variable (OS),
1246 always includes 'linux', 'mac' and 'win' in the set of valid values for OS,
1247 and allows conditions that depend on the set of all OSes, to the new format,
1248 which allows any set of variables, has no hardcoded values, and only allows
1249 explicit positive tests of variable values.
1251 conditions = value.get('conditions', [])
1252 if 'variables' not in value and all(len(cond) == 2 for cond in conditions):
1253 return value # Nothing to change
1255 def parse_condition(cond):
1256 return re.match(r'OS=="(\w+)"\Z', cond[0]).group(1)
1258 oses = set(map(parse_condition, conditions))
1259 default_oses = set(['linux', 'mac', 'win'])
1260 oses = sorted(oses | default_oses)
1262 def if_not_os(not_os, then):
1263 expr = ' or '.join('OS=="%s"' % os for os in oses if os != not_os)
1267 cond[:2] for cond in conditions if cond[1]
1269 if_not_os(parse_condition(cond), cond[2])
1270 for cond in conditions if len(cond) == 3
1273 if 'variables' in value:
1274 conditions.append(if_not_os(None, {'variables': value.pop('variables')}))
1277 value = value.copy()
1278 value['conditions'] = conditions
1282 def load_isolate_as_config(isolate_dir, value, file_comment):
1283 """Parses one .isolate file and returns a Configs() instance.
1285 |value| is the loaded dictionary that was defined in the gyp file.
1287 The expected format is strict, anything diverting from the format below will
1294 ['OS=="vms" and foo=42', {
1299 'isolate_dependency_tracked': [
1302 'isolate_dependency_untracked': [
1312 value = convert_old_to_new_format(value)
1314 variables_and_values = {}
1315 verify_root(value, variables_and_values)
1316 if variables_and_values:
1317 config_variables, config_values = zip(
1318 *sorted(variables_and_values.iteritems()))
1319 all_configs = list(itertools.product(*config_values))
1321 config_variables = None
1324 isolate = Configs(file_comment)
1326 # Add configuration-specific variables.
1327 for expr, then in value.get('conditions', []):
1328 configs = match_configs(expr, config_variables, all_configs)
1329 isolate.merge_dependencies(then['variables'], config_variables, configs)
1331 # Load the includes.
1332 for include in value.get('includes', []):
1333 if os.path.isabs(include):
1334 raise ExecutionError(
1335 'Failed to load configuration; absolute include path \'%s\'' %
1337 included_isolate = os.path.normpath(os.path.join(isolate_dir, include))
1338 with open(included_isolate, 'r') as f:
1339 included_isolate = load_isolate_as_config(
1340 os.path.dirname(included_isolate),
1341 eval_content(f.read()),
1343 isolate = union(isolate, included_isolate)
1348 def load_isolate_for_config(isolate_dir, content, variables):
1349 """Loads the .isolate file and returns the information unprocessed but
1350 filtered for the specific OS.
1352 Returns the command, dependencies and read_only flag. The dependencies are
1353 fixed to use os.path.sep.
1355 # Load the .isolate file, process its conditions, retrieve the command and
1357 isolate = load_isolate_as_config(isolate_dir, eval_content(content), None)
1359 config_name = tuple(variables[var] for var in isolate.config_variables)
1361 raise ExecutionError(
1362 'These configuration variables were missing from the command line: %s' %
1363 ', '.join(sorted(set(isolate.config_variables) - set(variables))))
1364 config = isolate.by_config.get(config_name)
1366 raise ExecutionError(
1367 'Failed to load configuration for variable \'%s\' for config(s) \'%s\''
1368 '\nAvailable configs: %s' %
1369 (', '.join(isolate.config_variables),
1370 ', '.join(config_name),
1371 ', '.join(str(s) for s in isolate.by_config)))
1372 # Merge tracked and untracked variables, isolate.py doesn't care about the
1373 # trackability of the variables, only the build tool does.
1375 f.replace('/', os.path.sep) for f in config.tracked + config.untracked
1377 touched = [f.replace('/', os.path.sep) for f in config.touched]
1378 return config.command, dependencies, touched, config.read_only
1381 def save_isolated(isolated, data):
1382 """Writes one or multiple .isolated files.
1384 Note: this reference implementation does not create child .isolated file so it
1385 always returns an empty list.
1387 Returns the list of child isolated files that are included by |isolated|.
1389 trace_inputs.write_json(isolated, data, True)
1393 def chromium_save_isolated(isolated, data, variables, algo):
1394 """Writes one or many .isolated files.
1396 This slightly increases the cold cache cost but greatly reduce the warm cache
1397 cost by splitting low-churn files off the master .isolated file. It also
1398 reduces overall isolateserver memcache consumption.
1402 def extract_into_included_isolated(prefix):
1404 'algo': data['algo'],
1407 'version': data['version'],
1409 for f in data['files'].keys():
1410 if f.startswith(prefix):
1411 new_slave['files'][f] = data['files'].pop(f)
1412 if new_slave['files']:
1413 slaves.append(new_slave)
1415 # Split test/data/ in its own .isolated file.
1416 extract_into_included_isolated(os.path.join('test', 'data', ''))
1418 # Split everything out of PRODUCT_DIR in its own .isolated file.
1419 if variables.get('PRODUCT_DIR'):
1420 extract_into_included_isolated(variables['PRODUCT_DIR'])
1423 for index, f in enumerate(slaves):
1424 slavepath = isolated[:-len('.isolated')] + '.%d.isolated' % index
1425 trace_inputs.write_json(slavepath, f, True)
1426 data.setdefault('includes', []).append(
1427 isolateserver.hash_file(slavepath, algo))
1428 files.append(os.path.basename(slavepath))
1430 files.extend(save_isolated(isolated, data))
1434 class Flattenable(object):
1435 """Represents data that can be represented as a json file."""
1439 """Returns a json-serializable version of itself.
1443 items = ((member, getattr(self, member)) for member in self.MEMBERS)
1444 return dict((member, value) for member, value in items if value is not None)
1447 def load(cls, data, *args, **kwargs):
1448 """Loads a flattened version."""
1450 out = cls(*args, **kwargs)
1451 for member in out.MEMBERS:
1453 # Access to a protected member XXX of a client class
1454 # pylint: disable=W0212
1455 out._load_member(member, data.pop(member))
1458 'Found unexpected entry %s while constructing an object %s' %
1459 (data, cls.__name__), data, cls.__name__)
1462 def _load_member(self, member, value):
1463 """Loads a member into self."""
1464 setattr(self, member, value)
1467 def load_file(cls, filename, *args, **kwargs):
1468 """Loads the data from a file or return an empty instance."""
1470 out = cls.load(trace_inputs.read_json(filename), *args, **kwargs)
1471 logging.debug('Loaded %s(%s)', cls.__name__, filename)
1472 except (IOError, ValueError) as e:
1473 # On failure, loads the default instance.
1474 out = cls(*args, **kwargs)
1475 logging.warn('Failed to load %s: %s', filename, e)
1479 class SavedState(Flattenable):
1480 """Describes the content of a .state file.
1482 This file caches the items calculated by this script and is used to increase
1483 the performance of the script. This file is not loaded by run_isolated.py.
1484 This file can always be safely removed.
1486 It is important to note that the 'files' dict keys are using native OS path
1487 separator instead of '/' used in .isolate file.
1490 # Algorithm used to generate the hash. The only supported value is at the
1491 # time of writting 'sha-1'.
1493 # Cache of the processed command. This value is saved because .isolated
1494 # files are never loaded by isolate.py so it's the only way to load the
1497 # Cache of the files found so the next run can skip hash calculation.
1499 # Path of the original .isolate file. Relative path to isolated_basedir.
1501 # List of included .isolated files. Used to support/remember 'slave'
1502 # .isolated files. Relative path to isolated_basedir.
1503 'child_isolated_files',
1504 # If the generated directory tree should be read-only.
1506 # Relative cwd to use to start the command.
1508 # GYP variables used to generate the .isolated file. Variables are saved so
1509 # a user can use isolate.py after building and the GYP variables are still
1512 # Version of the file format in format 'major.minor'. Any non-breaking
1513 # change must update minor. Any breaking change must update major.
1517 def __init__(self, isolated_basedir):
1518 """Creates an empty SavedState.
1520 |isolated_basedir| is the directory where the .isolated and .isolated.state
1523 super(SavedState, self).__init__()
1524 assert os.path.isabs(isolated_basedir), isolated_basedir
1525 assert os.path.isdir(isolated_basedir), isolated_basedir
1526 self.isolated_basedir = isolated_basedir
1528 # The default algorithm used.
1529 self.algo = isolateserver.SUPPORTED_ALGOS['sha-1']
1532 self.isolate_file = None
1533 self.child_isolated_files = []
1534 self.read_only = None
1535 self.relative_cwd = None
1536 self.variables = {'OS': get_flavor()}
1537 # The current version.
1538 self.version = '1.0'
1540 def update(self, isolate_file, variables):
1541 """Updates the saved state with new data to keep GYP variables and internal
1542 reference to the original .isolate file.
1544 assert os.path.isabs(isolate_file)
1545 # Convert back to a relative path. On Windows, if the isolate and
1546 # isolated files are on different drives, isolate_file will stay an absolute
1548 isolate_file = safe_relpath(isolate_file, self.isolated_basedir)
1550 # The same .isolate file should always be used to generate the .isolated and
1552 assert isolate_file == self.isolate_file or not self.isolate_file, (
1553 isolate_file, self.isolate_file)
1554 self.isolate_file = isolate_file
1555 self.variables.update(variables)
1557 def update_isolated(self, command, infiles, touched, read_only, relative_cwd):
1558 """Updates the saved state with data necessary to generate a .isolated file.
1560 The new files in |infiles| are added to self.files dict but their hash is
1561 not calculated here.
1563 self.command = command
1566 self.files.setdefault(f, {})
1568 self.files.setdefault(f, {})['T'] = True
1569 # Prune extraneous files that are not a dependency anymore.
1570 for f in set(self.files).difference(set(infiles).union(touched)):
1572 if read_only is not None:
1573 self.read_only = read_only
1574 self.relative_cwd = relative_cwd
1576 def to_isolated(self):
1577 """Creates a .isolated dictionary out of the saved state.
1579 https://code.google.com/p/swarming/wiki/IsolatedDesign
1582 """Returns a 'files' entry with only the whitelisted keys."""
1583 return dict((k, data[k]) for k in ('h', 'l', 'm', 's') if k in data)
1586 'algo': isolateserver.SUPPORTED_ALGOS_REVERSE[self.algo],
1588 (filepath, strip(data)) for filepath, data in self.files.iteritems()),
1589 'os': self.variables['OS'],
1590 'version': self.version,
1593 out['command'] = self.command
1594 if self.read_only is not None:
1595 out['read_only'] = self.read_only
1596 if self.relative_cwd:
1597 out['relative_cwd'] = self.relative_cwd
1601 def isolate_filepath(self):
1602 """Returns the absolute path of self.isolate_file."""
1603 return os.path.normpath(
1604 os.path.join(self.isolated_basedir, self.isolate_file))
1606 # Arguments number differs from overridden method
1608 def load(cls, data, isolated_basedir): # pylint: disable=W0221
1609 """Special case loading to disallow different OS.
1611 It is not possible to load a .isolated.state files from a different OS, this
1612 file is saved in OS-specific format.
1614 out = super(SavedState, cls).load(data, isolated_basedir)
1616 out.variables['OS'] = data['os']
1618 # Converts human readable form back into the proper class type.
1619 algo = data.get('algo', 'sha-1')
1620 if not algo in isolateserver.SUPPORTED_ALGOS:
1621 raise isolateserver.ConfigError('Unknown algo \'%s\'' % out.algo)
1622 out.algo = isolateserver.SUPPORTED_ALGOS[algo]
1624 # For example, 1.1 is guaranteed to be backward compatible with 1.0 code.
1625 if not re.match(r'^(\d+)\.(\d+)$', out.version):
1626 raise isolateserver.ConfigError('Unknown version \'%s\'' % out.version)
1627 if out.version.split('.', 1)[0] != '1':
1628 raise isolateserver.ConfigError(
1629 'Unsupported version \'%s\'' % out.version)
1631 # The .isolate file must be valid. It could be absolute on Windows if the
1632 # drive containing the .isolate and the drive containing the .isolated files
1634 assert not os.path.isabs(out.isolate_file) or sys.platform == 'win32'
1635 assert os.path.isfile(out.isolate_filepath), out.isolate_filepath
1639 """Makes sure 'algo' is in human readable form."""
1640 out = super(SavedState, self).flatten()
1641 out['algo'] = isolateserver.SUPPORTED_ALGOS_REVERSE[out['algo']]
1645 out = '%s(\n' % self.__class__.__name__
1646 out += ' command: %s\n' % self.command
1647 out += ' files: %d\n' % len(self.files)
1648 out += ' isolate_file: %s\n' % self.isolate_file
1649 out += ' read_only: %s\n' % self.read_only
1650 out += ' relative_cwd: %s\n' % self.relative_cwd
1651 out += ' child_isolated_files: %s\n' % self.child_isolated_files
1652 out += ' variables: %s' % ''.join(
1653 '\n %s=%s' % (k, self.variables[k]) for k in sorted(self.variables))
1658 class CompleteState(object):
1659 """Contains all the state to run the task at hand."""
1660 def __init__(self, isolated_filepath, saved_state):
1661 super(CompleteState, self).__init__()
1662 assert isolated_filepath is None or os.path.isabs(isolated_filepath)
1663 self.isolated_filepath = isolated_filepath
1664 # Contains the data to ease developer's use-case but that is not strictly
1666 self.saved_state = saved_state
1669 def load_files(cls, isolated_filepath):
1670 """Loads state from disk."""
1671 assert os.path.isabs(isolated_filepath), isolated_filepath
1672 isolated_basedir = os.path.dirname(isolated_filepath)
1675 SavedState.load_file(
1676 isolatedfile_to_state(isolated_filepath), isolated_basedir))
1678 def load_isolate(self, cwd, isolate_file, variables, ignore_broken_items):
1679 """Updates self.isolated and self.saved_state with information loaded from a
1682 Processes the loaded data, deduce root_dir, relative_cwd.
1684 # Make sure to not depend on os.getcwd().
1685 assert os.path.isabs(isolate_file), isolate_file
1686 isolate_file = file_path.get_native_path_case(isolate_file)
1688 'CompleteState.load_isolate(%s, %s, %s, %s)',
1689 cwd, isolate_file, variables, ignore_broken_items)
1690 relative_base_dir = os.path.dirname(isolate_file)
1692 # Processes the variables and update the saved state.
1693 variables = process_variables(cwd, variables, relative_base_dir)
1694 self.saved_state.update(isolate_file, variables)
1695 variables = self.saved_state.variables
1697 with open(isolate_file, 'r') as f:
1698 # At that point, variables are not replaced yet in command and infiles.
1699 # infiles may contain directory entries and is in posix style.
1700 command, infiles, touched, read_only = load_isolate_for_config(
1701 os.path.dirname(isolate_file), f.read(), variables)
1702 command = [eval_variables(i, variables) for i in command]
1703 infiles = [eval_variables(f, variables) for f in infiles]
1704 touched = [eval_variables(f, variables) for f in touched]
1705 # root_dir is automatically determined by the deepest root accessed with the
1706 # form '../../foo/bar'. Note that path variables must be taken in account
1707 # too, add them as if they were input files.
1708 path_variables = [variables[v] for v in PATH_VARIABLES if v in variables]
1709 root_dir = determine_root_dir(
1710 relative_base_dir, infiles + touched + path_variables)
1711 # The relative directory is automatically determined by the relative path
1712 # between root_dir and the directory containing the .isolate file,
1714 relative_cwd = os.path.relpath(relative_base_dir, root_dir)
1715 # Now that we know where the root is, check that the PATH_VARIABLES point
1717 for i in PATH_VARIABLES:
1719 if not path_starts_with(
1720 root_dir, os.path.join(relative_base_dir, variables[i])):
1721 raise isolateserver.MappingError(
1722 'Path variable %s=%r points outside the inferred root directory'
1723 ' %s' % (i, variables[i], root_dir))
1724 # Normalize the files based to root_dir. It is important to keep the
1725 # trailing os.path.sep at that step.
1727 relpath(normpath(os.path.join(relative_base_dir, f)), root_dir)
1731 relpath(normpath(os.path.join(relative_base_dir, f)), root_dir)
1734 follow_symlinks = variables['OS'] != 'win'
1735 # Expand the directories by listing each file inside. Up to now, trailing
1736 # os.path.sep must be kept. Do not expand 'touched'.
1737 infiles = expand_directories_and_symlinks(
1740 lambda x: re.match(r'.*\.(git|svn|pyc)$', x),
1742 ignore_broken_items)
1744 # If we ignore broken items then remove any missing touched items.
1745 if ignore_broken_items:
1746 original_touched_count = len(touched)
1747 touched = [touch for touch in touched if os.path.exists(touch)]
1749 if len(touched) != original_touched_count:
1750 logging.info('Removed %d invalid touched entries',
1751 len(touched) - original_touched_count)
1753 # Finally, update the new data to be able to generate the foo.isolated file,
1754 # the file that is used by run_isolated.py.
1755 self.saved_state.update_isolated(
1756 command, infiles, touched, read_only, relative_cwd)
1759 def process_inputs(self, subdir):
1760 """Updates self.saved_state.files with the files' mode and hash.
1762 If |subdir| is specified, filters to a subdirectory. The resulting .isolated
1765 See process_input() for more information.
1767 for infile in sorted(self.saved_state.files):
1768 if subdir and not infile.startswith(subdir):
1769 self.saved_state.files.pop(infile)
1771 filepath = os.path.join(self.root_dir, infile)
1772 self.saved_state.files[infile] = process_input(
1774 self.saved_state.files[infile],
1775 self.saved_state.read_only,
1776 self.saved_state.variables['OS'],
1777 self.saved_state.algo)
1779 def save_files(self):
1780 """Saves self.saved_state and creates a .isolated file."""
1781 logging.debug('Dumping to %s' % self.isolated_filepath)
1782 self.saved_state.child_isolated_files = chromium_save_isolated(
1783 self.isolated_filepath,
1784 self.saved_state.to_isolated(),
1785 self.saved_state.variables,
1786 self.saved_state.algo)
1788 i.get('s', 0) for i in self.saved_state.files.itervalues())
1790 # TODO(maruel): Stats are missing the .isolated files.
1791 logging.debug('Total size: %d bytes' % total_bytes)
1792 saved_state_file = isolatedfile_to_state(self.isolated_filepath)
1793 logging.debug('Dumping to %s' % saved_state_file)
1794 trace_inputs.write_json(saved_state_file, self.saved_state.flatten(), True)
1798 """Returns the absolute path of the root_dir to reference the .isolate file
1801 So that join(root_dir, relative_cwd, basename(isolate_file)) is equivalent
1802 to isolate_filepath.
1804 if not self.saved_state.isolate_file:
1805 raise ExecutionError('Please specify --isolate')
1806 isolate_dir = os.path.dirname(self.saved_state.isolate_filepath)
1808 if self.saved_state.relative_cwd == '.':
1809 root_dir = isolate_dir
1811 if not isolate_dir.endswith(self.saved_state.relative_cwd):
1812 raise ExecutionError(
1813 ('Make sure the .isolate file is in the directory that will be '
1814 'used as the relative directory. It is currently in %s and should '
1815 'be in %s') % (isolate_dir, self.saved_state.relative_cwd))
1816 # Walk back back to the root directory.
1817 root_dir = isolate_dir[:-(len(self.saved_state.relative_cwd) + 1)]
1818 return file_path.get_native_path_case(root_dir)
1821 def resultdir(self):
1822 """Returns the absolute path containing the .isolated file.
1824 It is usually equivalent to the variable PRODUCT_DIR. Uses the .isolated
1827 return os.path.dirname(self.isolated_filepath)
1830 def indent(data, indent_length):
1832 spacing = ' ' * indent_length
1833 return ''.join(spacing + l for l in str(data).splitlines(True))
1835 out = '%s(\n' % self.__class__.__name__
1836 out += ' root_dir: %s\n' % self.root_dir
1837 out += ' saved_state: %s)' % indent(self.saved_state, 2)
1841 def load_complete_state(options, cwd, subdir, skip_update):
1842 """Loads a CompleteState.
1844 This includes data from .isolate and .isolated.state files. Never reads the
1848 options: Options instance generated with OptionParserIsolate. For either
1849 options.isolate and options.isolated, if the value is set, it is an
1851 cwd: base directory to be used when loading the .isolate file.
1852 subdir: optional argument to only process file in the subdirectory, relative
1853 to CompleteState.root_dir.
1854 skip_update: Skip trying to load the .isolate file and processing the
1855 dependencies. It is useful when not needed, like when tracing.
1857 assert not options.isolate or os.path.isabs(options.isolate)
1858 assert not options.isolated or os.path.isabs(options.isolated)
1859 cwd = file_path.get_native_path_case(unicode(cwd))
1860 if options.isolated:
1861 # Load the previous state if it was present. Namely, "foo.isolated.state".
1862 # Note: this call doesn't load the .isolate file.
1863 complete_state = CompleteState.load_files(options.isolated)
1865 # Constructs a dummy object that cannot be saved. Useful for temporary
1866 # commands like 'run'.
1867 complete_state = CompleteState(None, SavedState())
1869 if not options.isolate:
1870 if not complete_state.saved_state.isolate_file:
1872 raise ExecutionError('A .isolate file is required.')
1875 isolate = complete_state.saved_state.isolate_filepath
1877 isolate = options.isolate
1878 if complete_state.saved_state.isolate_file:
1879 rel_isolate = safe_relpath(
1880 options.isolate, complete_state.saved_state.isolated_basedir)
1881 if rel_isolate != complete_state.saved_state.isolate_file:
1882 raise ExecutionError(
1883 '%s and %s do not match.' % (
1884 options.isolate, complete_state.saved_state.isolate_file))
1887 # Then load the .isolate and expands directories.
1888 complete_state.load_isolate(
1889 cwd, isolate, options.variables, options.ignore_broken_items)
1891 # Regenerate complete_state.saved_state.files.
1893 subdir = unicode(subdir)
1894 subdir = eval_variables(subdir, complete_state.saved_state.variables)
1895 subdir = subdir.replace('/', os.path.sep)
1898 complete_state.process_inputs(subdir)
1899 return complete_state
1902 def read_trace_as_isolate_dict(complete_state, trace_blacklist):
1903 """Reads a trace and returns the .isolate dictionary.
1905 Returns exceptions during the log parsing so it can be re-raised.
1907 api = trace_inputs.get_api()
1908 logfile = complete_state.isolated_filepath + '.log'
1909 if not os.path.isfile(logfile):
1910 raise ExecutionError(
1911 'No log file \'%s\' to read, did you forget to \'trace\'?' % logfile)
1913 data = api.parse_log(logfile, trace_blacklist, None)
1914 exceptions = [i['exception'] for i in data if 'exception' in i]
1915 results = (i['results'] for i in data if 'results' in i)
1916 results_stripped = (i.strip_root(complete_state.root_dir) for i in results)
1917 files = set(sum((result.existent for result in results_stripped), []))
1918 tracked, touched = split_touched(files)
1919 value = generate_isolate(
1923 complete_state.root_dir,
1924 complete_state.saved_state.variables,
1925 complete_state.saved_state.relative_cwd,
1927 return value, exceptions
1928 except trace_inputs.TracingFailure, e:
1929 raise ExecutionError(
1930 'Reading traces failed for: %s\n%s' %
1931 (' '.join(complete_state.saved_state.command), str(e)))
1934 def print_all(comment, data, stream):
1935 """Prints a complete .isolate file and its top-level file comment into a
1939 stream.write(comment)
1940 pretty_print(data, stream)
1943 def merge(complete_state, trace_blacklist):
1944 """Reads a trace and merges it back into the source .isolate file."""
1945 value, exceptions = read_trace_as_isolate_dict(
1946 complete_state, trace_blacklist)
1948 # Now take that data and union it into the original .isolate file.
1949 with open(complete_state.saved_state.isolate_filepath, 'r') as f:
1950 prev_content = f.read()
1951 isolate_dir = os.path.dirname(complete_state.saved_state.isolate_filepath)
1952 prev_config = load_isolate_as_config(
1954 eval_content(prev_content),
1955 extract_comment(prev_content))
1956 new_config = load_isolate_as_config(isolate_dir, value, '')
1957 config = union(prev_config, new_config)
1958 data = config.make_isolate_file()
1959 print('Updating %s' % complete_state.saved_state.isolate_file)
1960 with open(complete_state.saved_state.isolate_filepath, 'wb') as f:
1961 print_all(config.file_comment, data, f)
1963 # It got an exception, raise the first one.
1973 def CMDarchive(parser, args):
1974 """Creates a .isolated file and uploads the tree to an isolate server.
1976 All the files listed in the .isolated file are put in the isolate server
1977 cache via isolateserver.py.
1979 parser.add_option('--subdir', help='Filters to a subdirectory')
1980 options, args = parser.parse_args(args)
1982 parser.error('Unsupported argument: %s' % args)
1984 with tools.Profiler('GenerateHashtable'):
1987 complete_state = load_complete_state(
1988 options, os.getcwd(), options.subdir, False)
1989 if not options.outdir:
1990 options.outdir = os.path.join(
1991 os.path.dirname(complete_state.isolated_filepath), 'hashtable')
1992 # Make sure that complete_state isn't modified until save_files() is
1993 # called, because any changes made to it here will propagate to the files
1994 # created (which is probably not intended).
1995 complete_state.save_files()
1997 infiles = complete_state.saved_state.files
1998 # Add all the .isolated files.
2002 ] + complete_state.saved_state.child_isolated_files
2003 for item in isolated_files:
2004 item_path = os.path.join(
2005 os.path.dirname(complete_state.isolated_filepath), item)
2006 # Do not use isolateserver.hash_file() here because the file is
2007 # likely smallish (under 500kb) and its file size is needed.
2008 with open(item_path, 'rb') as f:
2010 isolated_hash.append(
2011 complete_state.saved_state.algo(content).hexdigest())
2012 isolated_metadata = {
2013 'h': isolated_hash[-1],
2017 infiles[item_path] = isolated_metadata
2019 logging.info('Creating content addressed object store with %d item',
2022 if is_url(options.outdir):
2023 isolateserver.upload_tree(
2024 base_url=options.outdir,
2025 indir=complete_state.root_dir,
2027 namespace='default-gzip')
2030 outdir=options.outdir,
2031 indir=complete_state.root_dir,
2033 action=run_isolated.HARDLINK_WITH_FALLBACK,
2036 print('%s %s' % (isolated_hash[0], os.path.basename(options.isolated)))
2038 # If the command failed, delete the .isolated file if it exists. This is
2039 # important so no stale swarm job is executed.
2040 if not success and os.path.isfile(options.isolated):
2041 os.remove(options.isolated)
2045 def CMDcheck(parser, args):
2046 """Checks that all the inputs are present and generates .isolated."""
2047 parser.add_option('--subdir', help='Filters to a subdirectory')
2048 options, args = parser.parse_args(args)
2050 parser.error('Unsupported argument: %s' % args)
2052 complete_state = load_complete_state(
2053 options, os.getcwd(), options.subdir, False)
2055 # Nothing is done specifically. Just store the result and state.
2056 complete_state.save_files()
2060 CMDhashtable = CMDarchive
2063 def CMDmerge(parser, args):
2064 """Reads and merges the data from the trace back into the original .isolate.
2068 parser.require_isolated = False
2069 add_trace_option(parser)
2070 options, args = parser.parse_args(args)
2072 parser.error('Unsupported argument: %s' % args)
2074 complete_state = load_complete_state(options, os.getcwd(), None, False)
2075 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist)
2076 merge(complete_state, blacklist)
2080 def CMDread(parser, args):
2081 """Reads the trace file generated with command 'trace'.
2085 parser.require_isolated = False
2086 add_trace_option(parser)
2088 '--skip-refresh', action='store_true',
2089 help='Skip reading .isolate file and do not refresh the hash of '
2092 '-m', '--merge', action='store_true',
2093 help='merge the results back in the .isolate file instead of printing')
2094 options, args = parser.parse_args(args)
2096 parser.error('Unsupported argument: %s' % args)
2098 complete_state = load_complete_state(
2099 options, os.getcwd(), None, options.skip_refresh)
2100 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist)
2101 value, exceptions = read_trace_as_isolate_dict(complete_state, blacklist)
2103 merge(complete_state, blacklist)
2105 pretty_print(value, sys.stdout)
2108 # It got an exception, raise the first one.
2116 def CMDremap(parser, args):
2117 """Creates a directory with all the dependencies mapped into it.
2119 Useful to test manually why a test is failing. The target executable is not
2122 parser.require_isolated = False
2123 options, args = parser.parse_args(args)
2125 parser.error('Unsupported argument: %s' % args)
2126 complete_state = load_complete_state(options, os.getcwd(), None, False)
2128 if not options.outdir:
2129 options.outdir = run_isolated.make_temp_dir(
2130 'isolate', complete_state.root_dir)
2132 if is_url(options.outdir):
2133 parser.error('Can\'t use url for --outdir with mode remap.')
2134 if not os.path.isdir(options.outdir):
2135 os.makedirs(options.outdir)
2136 print('Remapping into %s' % options.outdir)
2137 if len(os.listdir(options.outdir)):
2138 raise ExecutionError('Can\'t remap in a non-empty directory')
2140 outdir=options.outdir,
2141 indir=complete_state.root_dir,
2142 infiles=complete_state.saved_state.files,
2143 action=run_isolated.HARDLINK_WITH_FALLBACK,
2145 if complete_state.saved_state.read_only:
2146 run_isolated.make_writable(options.outdir, True)
2148 if complete_state.isolated_filepath:
2149 complete_state.save_files()
2153 def CMDrewrite(parser, args):
2154 """Rewrites a .isolate file into the canonical format."""
2155 parser.require_isolated = False
2156 options, args = parser.parse_args(args)
2158 parser.error('Unsupported argument: %s' % args)
2160 if options.isolated:
2161 # Load the previous state if it was present. Namely, "foo.isolated.state".
2162 complete_state = CompleteState.load_files(options.isolated)
2163 isolate = options.isolate or complete_state.saved_state.isolate_filepath
2165 isolate = options.isolate
2167 parser.error('--isolate is required.')
2169 with open(isolate, 'r') as f:
2171 config = load_isolate_as_config(
2172 os.path.dirname(os.path.abspath(isolate)),
2173 eval_content(content),
2174 extract_comment(content))
2175 data = config.make_isolate_file()
2176 print('Updating %s' % isolate)
2177 with open(isolate, 'wb') as f:
2178 print_all(config.file_comment, data, f)
2182 @subcommand.usage('-- [extra arguments]')
2183 def CMDrun(parser, args):
2184 """Runs the test executable in an isolated (temporary) directory.
2186 All the dependencies are mapped into the temporary directory and the
2187 directory is cleaned up after the target exits. Warning: if --outdir is
2188 specified, it is deleted upon exit.
2190 Argument processing stops at -- and these arguments are appended to the
2191 command line of the target to run. For example, use:
2192 isolate.py run --isolated foo.isolated -- --gtest_filter=Foo.Bar
2194 parser.require_isolated = False
2196 '--skip-refresh', action='store_true',
2197 help='Skip reading .isolate file and do not refresh the hash of '
2199 options, args = parser.parse_args(args)
2200 if options.outdir and is_url(options.outdir):
2201 parser.error('Can\'t use url for --outdir with mode run.')
2203 complete_state = load_complete_state(
2204 options, os.getcwd(), None, options.skip_refresh)
2205 cmd = complete_state.saved_state.command + args
2207 raise ExecutionError('No command to run.')
2209 cmd = tools.fix_python_path(cmd)
2211 root_dir = complete_state.root_dir
2212 if not options.outdir:
2213 if not os.path.isabs(root_dir):
2214 root_dir = os.path.join(os.path.dirname(options.isolated), root_dir)
2215 options.outdir = run_isolated.make_temp_dir('isolate', root_dir)
2217 if not os.path.isdir(options.outdir):
2218 os.makedirs(options.outdir)
2220 outdir=options.outdir,
2222 infiles=complete_state.saved_state.files,
2223 action=run_isolated.HARDLINK_WITH_FALLBACK,
2225 cwd = os.path.normpath(
2226 os.path.join(options.outdir, complete_state.saved_state.relative_cwd))
2227 if not os.path.isdir(cwd):
2228 # It can happen when no files are mapped from the directory containing the
2229 # .isolate file. But the directory must exist to be the current working
2232 if complete_state.saved_state.read_only:
2233 run_isolated.make_writable(options.outdir, True)
2234 logging.info('Running %s, cwd=%s' % (cmd, cwd))
2235 result = subprocess.call(cmd, cwd=cwd)
2238 run_isolated.rmtree(options.outdir)
2240 if complete_state.isolated_filepath:
2241 complete_state.save_files()
2245 @subcommand.usage('-- [extra arguments]')
2246 def CMDtrace(parser, args):
2247 """Traces the target using trace_inputs.py.
2249 It runs the executable without remapping it, and traces all the files it and
2250 its child processes access. Then the 'merge' command can be used to generate
2251 an updated .isolate file out of it or the 'read' command to print it out to
2254 Argument processing stops at -- and these arguments are appended to the
2255 command line of the target to run. For example, use:
2256 isolate.py trace --isolated foo.isolated -- --gtest_filter=Foo.Bar
2258 add_trace_option(parser)
2260 '-m', '--merge', action='store_true',
2261 help='After tracing, merge the results back in the .isolate file')
2263 '--skip-refresh', action='store_true',
2264 help='Skip reading .isolate file and do not refresh the hash of '
2266 options, args = parser.parse_args(args)
2268 complete_state = load_complete_state(
2269 options, os.getcwd(), None, options.skip_refresh)
2270 cmd = complete_state.saved_state.command + args
2272 raise ExecutionError('No command to run.')
2273 cmd = tools.fix_python_path(cmd)
2274 cwd = os.path.normpath(os.path.join(
2275 unicode(complete_state.root_dir),
2276 complete_state.saved_state.relative_cwd))
2277 cmd[0] = os.path.normpath(os.path.join(cwd, cmd[0]))
2278 if not os.path.isfile(cmd[0]):
2279 raise ExecutionError(
2280 'Tracing failed for: %s\nIt doesn\'t exit' % ' '.join(cmd))
2281 logging.info('Running %s, cwd=%s' % (cmd, cwd))
2282 api = trace_inputs.get_api()
2283 logfile = complete_state.isolated_filepath + '.log'
2284 api.clean_trace(logfile)
2287 with api.get_tracer(logfile) as tracer:
2288 result, out = tracer.trace(
2293 except trace_inputs.TracingFailure, e:
2294 raise ExecutionError('Tracing failed for: %s\n%s' % (' '.join(cmd), str(e)))
2298 'Tracer exited with %d, which means the tests probably failed so the '
2299 'trace is probably incomplete.', result)
2302 complete_state.save_files()
2305 blacklist = trace_inputs.gen_blacklist(options.trace_blacklist)
2306 merge(complete_state, blacklist)
2311 def _process_variable_arg(_option, _opt, _value, parser):
2312 if not parser.rargs:
2313 raise optparse.OptionValueError(
2314 'Please use --variable FOO=BAR or --variable FOO BAR')
2315 k = parser.rargs.pop(0)
2317 parser.values.variables.append(tuple(k.split('=', 1)))
2319 if not parser.rargs:
2320 raise optparse.OptionValueError(
2321 'Please use --variable FOO=BAR or --variable FOO BAR')
2322 v = parser.rargs.pop(0)
2323 parser.values.variables.append((k, v))
2326 def add_variable_option(parser):
2327 """Adds --isolated and --variable to an OptionParser."""
2331 help='.isolated file to generate or read')
2332 # Keep for compatibility. TODO(maruel): Remove once not used anymore.
2336 help=optparse.SUPPRESS_HELP)
2337 default_variables = [('OS', get_flavor())]
2338 if sys.platform in ('win32', 'cygwin'):
2339 default_variables.append(('EXECUTABLE_SUFFIX', '.exe'))
2341 default_variables.append(('EXECUTABLE_SUFFIX', ''))
2345 callback=_process_variable_arg,
2346 default=default_variables,
2349 help='Variables to process in the .isolate file, default: %default. '
2350 'Variables are persistent accross calls, they are saved inside '
2351 '<.isolated>.state')
2354 def add_trace_option(parser):
2355 """Adds --trace-blacklist to the parser."""
2357 '--trace-blacklist',
2358 action='append', default=list(DEFAULT_BLACKLIST),
2359 help='List of regexp to use as blacklist filter for files to consider '
2360 'important, not to be confused with --blacklist which blacklists '
2364 def parse_isolated_option(parser, options, cwd, require_isolated):
2365 """Processes --isolated."""
2366 if options.isolated:
2367 options.isolated = os.path.normpath(
2368 os.path.join(cwd, options.isolated.replace('/', os.path.sep)))
2369 if require_isolated and not options.isolated:
2370 parser.error('--isolated is required.')
2371 if options.isolated and not options.isolated.endswith('.isolated'):
2372 parser.error('--isolated value must end with \'.isolated\'')
2375 def parse_variable_option(options):
2376 """Processes --variable."""
2377 # TODO(benrg): Maybe we should use a copy of gyp's NameValueListToDict here,
2378 # but it wouldn't be backward compatible.
2379 def try_make_int(s):
2380 """Converts a value to int if possible, converts to unicode otherwise."""
2384 return s.decode('utf-8')
2385 options.variables = dict((k, try_make_int(v)) for k, v in options.variables)
2388 class OptionParserIsolate(tools.OptionParserWithLogging):
2389 """Adds automatic --isolate, --isolated, --out and --variable handling."""
2390 # Set it to False if it is not required, e.g. it can be passed on but do not
2391 # fail if not given.
2392 require_isolated = True
2394 def __init__(self, **kwargs):
2395 tools.OptionParserWithLogging.__init__(
2397 verbose=int(os.environ.get('ISOLATE_DEBUG', 0)),
2399 group = optparse.OptionGroup(self, "Common options")
2403 help='.isolate file to load the dependency data from')
2404 add_variable_option(group)
2406 '-o', '--outdir', metavar='DIR',
2407 help='Directory used to recreate the tree or store the hash table. '
2408 'Defaults: run|remap: a /tmp subdirectory, others: '
2409 'defaults to the directory containing --isolated')
2411 '--ignore_broken_items', action='store_true',
2412 default=bool(os.environ.get('ISOLATE_IGNORE_BROKEN_ITEMS')),
2413 help='Indicates that invalid entries in the isolated file to be '
2414 'only be logged and not stop processing. Defaults to True if '
2415 'env var ISOLATE_IGNORE_BROKEN_ITEMS is set')
2416 self.add_option_group(group)
2418 def parse_args(self, *args, **kwargs):
2419 """Makes sure the paths make sense.
2421 On Windows, / and \ are often mixed together in a path.
2423 options, args = tools.OptionParserWithLogging.parse_args(
2424 self, *args, **kwargs)
2425 if not self.allow_interspersed_args and args:
2426 self.error('Unsupported argument: %s' % args)
2428 cwd = file_path.get_native_path_case(unicode(os.getcwd()))
2429 parse_isolated_option(self, options, cwd, self.require_isolated)
2430 parse_variable_option(options)
2433 # TODO(maruel): Work with non-ASCII.
2434 # The path must be in native path case for tracing purposes.
2435 options.isolate = unicode(options.isolate).replace('/', os.path.sep)
2436 options.isolate = os.path.normpath(os.path.join(cwd, options.isolate))
2437 options.isolate = file_path.get_native_path_case(options.isolate)
2439 if options.outdir and not is_url(options.outdir):
2440 options.outdir = unicode(options.outdir).replace('/', os.path.sep)
2441 # outdir doesn't need native path case since tracing is never done from
2443 options.outdir = os.path.normpath(os.path.join(cwd, options.outdir))
2445 return options, args
2449 dispatcher = subcommand.CommandDispatcher(__name__)
2451 return dispatcher.execute(OptionParserIsolate(version=__version__), argv)
2454 isolateserver.ConfigError,
2455 isolateserver.MappingError) as e:
2456 sys.stderr.write('\nError: ')
2457 sys.stderr.write(str(e))
2458 sys.stderr.write('\n')
2462 if __name__ == '__main__':
2463 fix_encoding.fix_encoding()
2464 tools.disable_buffering()
2466 sys.exit(main(sys.argv[1:]))