2 # Copyright 2014 The Chromium Authors. All rights reserved.
3 # Use of this source code is governed by a BSD-style license that can be
4 # found in the LICENSE file.
6 """A utility script for downloading versioned Syzygy binaries."""
24 _LOGGER
= logging
.getLogger(os
.path
.basename(__file__
))
26 # The relative path where official builds are archived in their GS bucket.
27 _SYZYGY_ARCHIVE_PATH
= ('/builds/official/%(revision)s')
29 # A JSON file containing the state of the download directory. If this file and
30 # directory state do not agree, then the binaries will be downloaded and
34 # This matches an integer (an SVN revision number) or a SHA1 value (a GIT hash).
35 # The archive exclusively uses lowercase GIT hashes.
36 _REVISION_RE
= re
.compile('^(?:\d+|[a-f0-9]{40})$')
38 # This matches an MD5 hash.
39 _MD5_RE
= re
.compile('^[a-f0-9]{32}$')
41 # List of reources to be downloaded and installed. These are tuples with the
43 # (basename, logging name, relative installation path, extraction filter)
45 ('benchmark.zip', 'benchmark', '', None),
46 ('binaries.zip', 'binaries', 'exe', None),
47 ('symbols.zip', 'symbols', 'exe',
48 lambda x
: x
.filename
.endswith('.dll.pdb'))]
51 def _LoadState(output_dir
):
52 """Loads the contents of the state file for a given |output_dir|, returning
53 None if it doesn't exist.
55 path
= os
.path
.join(output_dir
, _STATE
)
56 if not os
.path
.exists(path
):
57 _LOGGER
.debug('No state file found.')
59 with
open(path
, 'rb') as f
:
60 _LOGGER
.debug('Reading state file: %s', path
)
64 _LOGGER
.debug('Invalid state file.')
68 def _SaveState(output_dir
, state
, dry_run
=False):
69 """Saves the |state| dictionary to the given |output_dir| as a JSON file."""
70 path
= os
.path
.join(output_dir
, _STATE
)
71 _LOGGER
.debug('Writing state file: %s', path
)
74 with
open(path
, 'wb') as f
:
75 f
.write(json
.dumps(state
, sort_keys
=True, indent
=2))
79 """Returns the MD5 hash of the file at |path|, which must exist."""
80 return hashlib
.md5(open(path
, 'rb').read()).hexdigest()
83 def _StateIsValid(state
):
84 """Returns true if the given state structure is valid."""
85 if not isinstance(state
, dict):
86 _LOGGER
.debug('State must be a dict.')
88 r
= state
.get('revision', None)
89 if not isinstance(r
, basestring
) or not _REVISION_RE
.match(r
):
90 _LOGGER
.debug('State contains an invalid revision.')
92 c
= state
.get('contents', None)
93 if not isinstance(c
, dict):
94 _LOGGER
.debug('State must contain a contents dict.')
96 for (relpath
, md5
) in c
.iteritems():
97 if not isinstance(relpath
, basestring
) or len(relpath
) == 0:
98 _LOGGER
.debug('State contents dict contains an invalid path.')
100 if not isinstance(md5
, basestring
) or not _MD5_RE
.match(md5
):
101 _LOGGER
.debug('State contents dict contains an invalid MD5 digest.')
106 def _BuildActualState(stored
, revision
, output_dir
):
107 """Builds the actual state using the provided |stored| state as a template.
108 Only examines files listed in the stored state, causing the script to ignore
109 files that have been added to the directories locally. |stored| must be a
110 valid state dictionary.
113 state
= { 'revision': revision
, 'contents': contents
}
114 for relpath
, md5
in stored
['contents'].iteritems():
115 abspath
= os
.path
.abspath(os
.path
.join(output_dir
, relpath
))
116 if os
.path
.isfile(abspath
):
118 contents
[relpath
] = m
123 def _StatesAreConsistent(stored
, actual
):
124 """Validates whether two state dictionaries are consistent. Both must be valid
125 state dictionaries. Additional entries in |actual| are ignored.
127 if stored
['revision'] != actual
['revision']:
128 _LOGGER
.debug('Mismatched revision number.')
130 cont_stored
= stored
['contents']
131 cont_actual
= actual
['contents']
132 for relpath
, md5
in cont_stored
.iteritems():
133 if relpath
not in cont_actual
:
134 _LOGGER
.debug('Missing content: %s', relpath
)
136 if md5
!= cont_actual
[relpath
]:
137 _LOGGER
.debug('Modified content: %s', relpath
)
142 def _GetCurrentState(revision
, output_dir
):
143 """Loads the current state and checks to see if it is consistent. Returns
144 a tuple (state, bool). The returned state will always be valid, even if an
145 invalid state is present on disk.
147 stored
= _LoadState(output_dir
)
148 if not _StateIsValid(stored
):
149 _LOGGER
.debug('State is invalid.')
150 # Return a valid but empty state.
151 return ({'revision': '0', 'contents': {}}, False)
152 actual
= _BuildActualState(stored
, revision
, output_dir
)
153 # If the script has been modified consider the state invalid.
154 path
= os
.path
.join(output_dir
, _STATE
)
155 if os
.path
.getmtime(__file__
) > os
.path
.getmtime(path
):
156 return (stored
, False)
157 # Otherwise, explicitly validate the state.
158 if not _StatesAreConsistent(stored
, actual
):
159 return (stored
, False)
160 return (stored
, True)
163 def _DirIsEmpty(path
):
164 """Returns true if the given directory is empty, false otherwise."""
165 for root
, dirs
, files
in os
.walk(path
):
166 return not dirs
and not files
169 def _RmTreeHandleReadOnly(func
, path
, exc
):
170 """An error handling function for use with shutil.rmtree. This will
171 detect failures to remove read-only files, and will change their properties
172 prior to removing them. This is necessary on Windows as os.remove will return
173 an access error for read-only files, and git repos contain read-only
177 if func
in (os
.rmdir
, os
.remove
) and excvalue
.errno
== errno
.EACCES
:
178 _LOGGER
.debug('Removing read-only path: %s', path
)
179 os
.chmod(path
, stat
.S_IRWXU | stat
.S_IRWXG | stat
.S_IRWXO
)
186 """A wrapper of shutil.rmtree that handles read-only files."""
187 shutil
.rmtree(path
, ignore_errors
=False, onerror
=_RmTreeHandleReadOnly
)
190 def _CleanState(output_dir
, state
, dry_run
=False):
191 """Cleans up files/directories in |output_dir| that are referenced by
192 the given |state|. Raises an error if there are local changes. Returns a
193 dictionary of files that were deleted.
195 _LOGGER
.debug('Deleting files from previous installation.')
198 # Generate a list of files to delete, relative to |output_dir|.
199 contents
= state
['contents']
200 files
= sorted(contents
.keys())
202 # Try to delete the files. Keep track of directories to delete as well.
204 for relpath
in files
:
205 fullpath
= os
.path
.join(output_dir
, relpath
)
206 fulldir
= os
.path
.dirname(fullpath
)
208 if os
.path
.exists(fullpath
):
209 # If somehow the file has become a directory complain about it.
210 if os
.path
.isdir(fullpath
):
211 raise Exception('Directory exists where file expected: %s' % fullpath
)
213 # Double check that the file doesn't have local changes. If it does
214 # then refuse to delete it.
215 if relpath
in contents
:
216 stored_md5
= contents
[relpath
]
217 actual_md5
= _Md5(fullpath
)
218 if actual_md5
!= stored_md5
:
219 raise Exception('File has local changes: %s' % fullpath
)
221 # The file is unchanged so it can safely be deleted.
222 _LOGGER
.debug('Deleting file "%s".', fullpath
)
223 deleted
[relpath
] = True
227 # Sort directories from longest name to shortest. This lets us remove empty
228 # directories from the most nested paths first.
229 dirs
= sorted(dirs
.keys(), key
=lambda x
: len(x
), reverse
=True)
231 if os
.path
.exists(p
) and _DirIsEmpty(p
):
232 _LOGGER
.debug('Deleting empty directory "%s".', p
)
240 """Looks for depot_tools and returns the absolute path to gsutil.py."""
241 for path
in os
.environ
['PATH'].split(os
.pathsep
):
242 path
= os
.path
.abspath(path
)
243 git_cl
= os
.path
.join(path
, 'git_cl.py')
244 gs_util
= os
.path
.join(path
, 'gsutil.py')
245 if os
.path
.exists(git_cl
) and os
.path
.exists(gs_util
):
251 """Runs the given command in gsutil with exponential backoff and retries."""
252 gs_util
= _FindGsUtil()
253 cmd
= [sys
.executable
, gs_util
] + list(cmd
)
256 timeout
= 4 # Seconds.
258 _LOGGER
.debug('Running %s', cmd
)
259 prog
= subprocess
.Popen(cmd
, shell
=False)
262 # Stop retrying on success.
263 if prog
.returncode
== 0:
266 # Raise a permanent failure if retries have been exhausted.
268 raise RuntimeError('Command "%s" returned %d.' % (cmd
, prog
.returncode
))
270 _LOGGER
.debug('Sleeping %d seconds and trying again.', timeout
)
276 def _Download(resource
):
277 """Downloads the given GS resource to a temporary file, returning its path."""
278 tmp
= tempfile
.mkstemp(suffix
='syzygy_archive')
280 url
= 'gs://syzygy-archive' + resource
281 _GsUtil('cp', url
, tmp
[1])
285 def _InstallBinaries(options
, deleted
={}):
286 """Installs Syzygy binaries. This assumes that the output directory has
287 already been cleaned, as it will refuse to overwrite existing files."""
289 state
= { 'revision': options
.revision
, 'contents': contents
}
290 archive_path
= _SYZYGY_ARCHIVE_PATH
% { 'revision': options
.revision
}
291 if options
.resources
:
292 resources
= [(resource
, resource
, '', None)
293 for resource
in options
.resources
]
295 resources
= _RESOURCES
296 for (base
, name
, subdir
, filt
) in resources
:
297 # Create the output directory if it doesn't exist.
298 fulldir
= os
.path
.join(options
.output_dir
, subdir
)
299 if os
.path
.isfile(fulldir
):
300 raise Exception('File exists where a directory needs to be created: %s' %
302 if not os
.path
.exists(fulldir
):
303 _LOGGER
.debug('Creating directory: %s', fulldir
)
304 if not options
.dry_run
:
307 # Download and read the archive.
308 resource
= archive_path
+ '/' + base
309 _LOGGER
.debug('Retrieving %s archive at "%s".', name
, resource
)
310 path
= _Download(resource
)
312 _LOGGER
.debug('Unzipping %s archive.', name
)
313 with
open(path
, 'rb') as data
:
314 archive
= zipfile
.ZipFile(data
)
315 for entry
in archive
.infolist():
316 if not filt
or filt(entry
):
317 fullpath
= os
.path
.normpath(os
.path
.join(fulldir
, entry
.filename
))
318 relpath
= os
.path
.relpath(fullpath
, options
.output_dir
)
319 if os
.path
.exists(fullpath
):
320 # If in a dry-run take into account the fact that the file *would*
322 if options
.dry_run
and relpath
in deleted
:
325 raise Exception('Path already exists: %s' % fullpath
)
327 # Extract the file and update the state dictionary.
328 _LOGGER
.debug('Extracting "%s".', fullpath
)
329 if not options
.dry_run
:
330 archive
.extract(entry
.filename
, fulldir
)
332 contents
[relpath
] = md5
333 if sys
.platform
== 'cygwin':
334 os
.chmod(fullpath
, os
.stat(fullpath
).st_mode | stat
.S_IXUSR
)
336 _LOGGER
.debug('Removing temporary file "%s".', path
)
342 def _ParseCommandLine():
343 """Parses the command-line and returns an options structure."""
344 option_parser
= optparse
.OptionParser()
345 option_parser
.add_option('--dry-run', action
='store_true', default
=False,
346 help='If true then will simply list actions that would be performed.')
347 option_parser
.add_option('--force', action
='store_true', default
=False,
348 help='Force an installation even if the binaries are up to date.')
349 option_parser
.add_option('--no-cleanup', action
='store_true', default
=False,
350 help='Allow installation on non-Windows platforms, and skip the forced '
352 option_parser
.add_option('--output-dir', type='string',
353 help='The path where the binaries will be replaced. Existing binaries '
354 'will only be overwritten if not up to date.')
355 option_parser
.add_option('--overwrite', action
='store_true', default
=False,
356 help='If specified then the installation will happily delete and rewrite '
357 'the entire output directory, blasting any local changes.')
358 option_parser
.add_option('--revision', type='string',
359 help='The SVN revision or GIT hash associated with the required version.')
360 option_parser
.add_option('--revision-file', type='string',
361 help='A text file containing an SVN revision or GIT hash.')
362 option_parser
.add_option('--resource', type='string', action
='append',
363 dest
='resources', help='A resource to be downloaded.')
364 option_parser
.add_option('--verbose', dest
='log_level', action
='store_const',
365 default
=logging
.INFO
, const
=logging
.DEBUG
,
366 help='Enables verbose logging.')
367 option_parser
.add_option('--quiet', dest
='log_level', action
='store_const',
368 default
=logging
.INFO
, const
=logging
.ERROR
,
369 help='Disables all output except for errors.')
370 options
, args
= option_parser
.parse_args()
372 option_parser
.error('Unexpected arguments: %s' % args
)
373 if not options
.output_dir
:
374 option_parser
.error('Must specify --output-dir.')
375 if not options
.revision
and not options
.revision_file
:
376 option_parser
.error('Must specify one of --revision or --revision-file.')
377 if options
.revision
and options
.revision_file
:
378 option_parser
.error('Must not specify both --revision and --revision-file.')
381 logging
.basicConfig(level
=options
.log_level
)
383 # If a revision file has been specified then read it.
384 if options
.revision_file
:
385 options
.revision
= open(options
.revision_file
, 'rb').read().strip()
386 _LOGGER
.debug('Parsed revision "%s" from file "%s".',
387 options
.revision
, options
.revision_file
)
389 # Ensure that the specified SVN revision or GIT hash is valid.
390 if not _REVISION_RE
.match(options
.revision
):
391 option_parser
.error('Must specify a valid SVN or GIT revision.')
393 # This just makes output prettier to read.
394 options
.output_dir
= os
.path
.normpath(options
.output_dir
)
399 def _RemoveOrphanedFiles(options
):
400 """This is run on non-Windows systems to remove orphaned files that may have
401 been downloaded by a previous version of this script.
403 # Reconfigure logging to output info messages. This will allow inspection of
404 # cleanup status on non-Windows buildbots.
405 _LOGGER
.setLevel(logging
.INFO
)
407 output_dir
= os
.path
.abspath(options
.output_dir
)
409 # We only want to clean up the folder in 'src/third_party/syzygy', and we
410 # expect to be called with that as an output directory. This is an attempt to
411 # not start deleting random things if the script is run from an alternate
412 # location, or not called from the gclient hooks.
413 expected_syzygy_dir
= os
.path
.abspath(os
.path
.join(
414 os
.path
.dirname(__file__
), '..', 'third_party', 'syzygy'))
415 expected_output_dir
= os
.path
.join(expected_syzygy_dir
, 'binaries')
416 if expected_output_dir
!= output_dir
:
417 _LOGGER
.info('Unexpected output directory, skipping cleanup.')
420 if not os
.path
.isdir(expected_syzygy_dir
):
421 _LOGGER
.info('Output directory does not exist, skipping cleanup.')
424 def OnError(function
, path
, excinfo
):
425 """Logs error encountered by shutil.rmtree."""
426 _LOGGER
.error('Error when running %s(%s)', function
, path
, exc_info
=excinfo
)
428 _LOGGER
.info('Removing orphaned files from %s', expected_syzygy_dir
)
429 if not options
.dry_run
:
430 shutil
.rmtree(expected_syzygy_dir
, True, OnError
)
434 options
= _ParseCommandLine()
437 _LOGGER
.debug('Performing a dry-run.')
439 # We only care about Windows platforms, as the Syzygy binaries aren't used
440 # elsewhere. However, there was a short period of time where this script
441 # wasn't gated on OS types, and those OSes downloaded and installed binaries.
442 # This will cleanup orphaned files on those operating systems.
443 if sys
.platform
not in ('win32', 'cygwin'):
444 if options
.no_cleanup
:
445 _LOGGER
.debug('Skipping usual cleanup for non-Windows platforms.')
447 return _RemoveOrphanedFiles(options
)
449 # Load the current installation state, and validate it against the
450 # requested installation.
451 state
, is_consistent
= _GetCurrentState(options
.revision
, options
.output_dir
)
453 # Decide whether or not an install is necessary.
455 _LOGGER
.debug('Forcing reinstall of binaries.')
457 # Avoid doing any work if the contents of the directory are consistent.
458 _LOGGER
.debug('State unchanged, no reinstall necessary.')
461 # Under normal logging this is the only only message that will be reported.
462 _LOGGER
.info('Installing revision %s Syzygy binaries.',
463 options
.revision
[0:12])
465 # Clean up the old state to begin with.
467 if options
.overwrite
:
468 if os
.path
.exists(options
.output_dir
):
469 # If overwrite was specified then take a heavy-handed approach.
470 _LOGGER
.debug('Deleting entire installation directory.')
471 if not options
.dry_run
:
472 _RmTree(options
.output_dir
)
474 # Otherwise only delete things that the previous installation put in place,
475 # and take care to preserve any local changes.
476 deleted
= _CleanState(options
.output_dir
, state
, options
.dry_run
)
478 # Install the new binaries. In a dry-run this will actually download the
479 # archives, but it won't write anything to disk.
480 state
= _InstallBinaries(options
, deleted
)
482 # Build and save the state for the directory.
483 _SaveState(options
.output_dir
, state
, options
.dry_run
)
486 if __name__
== '__main__':