2 # Copyright 2014 The Chromium Authors. All rights reserved.
3 # Use of this source code is governed by a BSD-style license that can be
4 # found in the LICENSE file.
6 """Generate a spatial analysis against an arbitrary library.
8 To use, build the 'binary_size_tool' target. Then run this tool, passing
9 in the location of the library to be analyzed along with any other options
16 import multiprocessing
27 import binary_size_utils
29 # This path change is not beautiful. Temporary (I hope) measure until
30 # the chromium project has figured out a proper way to organize the
31 # library of python tools. http://crbug.com/375725
32 elf_symbolizer_path
= os
.path
.abspath(os
.path
.join(
33 os
.path
.dirname(__file__
),
39 sys
.path
.append(elf_symbolizer_path
)
40 import symbols
.elf_symbolizer
as elf_symbolizer
# pylint: disable=F0401
43 # Node dictionary keys. These are output in json read by the webapp so
44 # keep them short to save file size.
45 # Note: If these change, the webapp must also change.
48 NODE_CHILDREN_KEY
= 'children'
49 NODE_SYMBOL_TYPE_KEY
= 't'
50 NODE_SYMBOL_SIZE_KEY
= 'value'
51 NODE_MAX_DEPTH_KEY
= 'maxDepth'
52 NODE_LAST_PATH_ELEMENT_KEY
= 'lastPathElement'
54 # The display name of the bucket where we put symbols without path.
55 NAME_NO_PATH_BUCKET
= '(No Path)'
57 # Try to keep data buckets smaller than this to avoid killing the
59 BIG_BUCKET_LIMIT
= 3000
62 def _MkChild(node
, name
):
63 child
= node
[NODE_CHILDREN_KEY
].get(name
)
65 child
= {NODE_NAME_KEY
: name
,
66 NODE_CHILDREN_KEY
: {}}
67 node
[NODE_CHILDREN_KEY
][name
] = child
72 def SplitNoPathBucket(node
):
73 """NAME_NO_PATH_BUCKET can be too large for the graphing lib to
74 handle. Split it into sub-buckets in that case."""
75 root_children
= node
[NODE_CHILDREN_KEY
]
76 if NAME_NO_PATH_BUCKET
in root_children
:
77 no_path_bucket
= root_children
[NAME_NO_PATH_BUCKET
]
78 old_children
= no_path_bucket
[NODE_CHILDREN_KEY
]
80 for symbol_type
, symbol_bucket
in old_children
.iteritems():
81 count
+= len(symbol_bucket
[NODE_CHILDREN_KEY
])
82 if count
> BIG_BUCKET_LIMIT
:
84 no_path_bucket
[NODE_CHILDREN_KEY
] = new_children
87 for symbol_type
, symbol_bucket
in old_children
.iteritems():
88 for symbol_name
, value
in symbol_bucket
[NODE_CHILDREN_KEY
].iteritems():
89 if index
% BIG_BUCKET_LIMIT
== 0:
90 group_no
= (index
/ BIG_BUCKET_LIMIT
) + 1
91 current_bucket
= _MkChild(no_path_bucket
,
92 '%s subgroup %d' % (NAME_NO_PATH_BUCKET
,
94 assert not NODE_TYPE_KEY
in node
or node
[NODE_TYPE_KEY
] == 'p'
95 node
[NODE_TYPE_KEY
] = 'p' # p for path
97 symbol_size
= value
[NODE_SYMBOL_SIZE_KEY
]
98 AddSymbolIntoFileNode(current_bucket
, symbol_type
,
99 symbol_name
, symbol_size
)
102 def MakeChildrenDictsIntoLists(node
):
104 if NODE_CHILDREN_KEY
in node
:
105 largest_list_len
= len(node
[NODE_CHILDREN_KEY
])
107 for child
in node
[NODE_CHILDREN_KEY
].itervalues():
108 child_largest_list_len
= MakeChildrenDictsIntoLists(child
)
109 if child_largest_list_len
> largest_list_len
:
110 largest_list_len
= child_largest_list_len
111 child_list
.append(child
)
112 node
[NODE_CHILDREN_KEY
] = child_list
114 return largest_list_len
117 def AddSymbolIntoFileNode(node
, symbol_type
, symbol_name
, symbol_size
):
118 """Puts symbol into the file path node |node|.
119 Returns the number of added levels in tree. I.e. returns 2."""
121 # 'node' is the file node and first step is to find its symbol-type bucket.
122 node
[NODE_LAST_PATH_ELEMENT_KEY
] = True
123 node
= _MkChild(node
, symbol_type
)
124 assert not NODE_TYPE_KEY
in node
or node
[NODE_TYPE_KEY
] == 'b'
125 node
[NODE_SYMBOL_TYPE_KEY
] = symbol_type
126 node
[NODE_TYPE_KEY
] = 'b' # b for bucket
128 # 'node' is now the symbol-type bucket. Make the child entry.
129 node
= _MkChild(node
, symbol_name
)
130 if NODE_CHILDREN_KEY
in node
:
131 if node
[NODE_CHILDREN_KEY
]:
132 logging
.warning('A container node used as symbol for %s.' % symbol_name
)
133 # This is going to be used as a leaf so no use for child list.
134 del node
[NODE_CHILDREN_KEY
]
135 node
[NODE_SYMBOL_SIZE_KEY
] = symbol_size
136 node
[NODE_SYMBOL_TYPE_KEY
] = symbol_type
137 node
[NODE_TYPE_KEY
] = 's' # s for symbol
139 return 2 # Depth of the added subtree.
142 def MakeCompactTree(symbols
, symbol_path_origin_dir
):
143 result
= {NODE_NAME_KEY
: '/',
144 NODE_CHILDREN_KEY
: {},
146 NODE_MAX_DEPTH_KEY
: 0}
147 seen_symbol_with_path
= False
148 cwd
= os
.path
.abspath(os
.getcwd())
149 for symbol_name
, symbol_type
, symbol_size
, file_path
, _address
in symbols
:
151 if 'vtable for ' in symbol_name
:
152 symbol_type
= '@' # hack to categorize these separately
153 # Take path like '/foo/bar/baz', convert to ['foo', 'bar', 'baz']
154 if file_path
and file_path
!= "??":
155 file_path
= os
.path
.abspath(os
.path
.join(symbol_path_origin_dir
,
157 # Let the output structure be relative to $CWD if inside $CWD,
158 # otherwise relative to the disk root. This is to avoid
159 # unnecessary click-through levels in the output.
160 if file_path
.startswith(cwd
+ os
.sep
):
161 file_path
= file_path
[len(cwd
):]
162 if file_path
.startswith('/'):
163 file_path
= file_path
[1:]
164 seen_symbol_with_path
= True
166 file_path
= NAME_NO_PATH_BUCKET
168 path_parts
= file_path
.split('/')
170 # Find pre-existing node in tree, or update if it already exists
173 while len(path_parts
) > 0:
174 path_part
= path_parts
.pop(0)
175 if len(path_part
) == 0:
178 node
= _MkChild(node
, path_part
)
179 assert not NODE_TYPE_KEY
in node
or node
[NODE_TYPE_KEY
] == 'p'
180 node
[NODE_TYPE_KEY
] = 'p' # p for path
182 depth
+= AddSymbolIntoFileNode(node
, symbol_type
, symbol_name
, symbol_size
)
183 result
[NODE_MAX_DEPTH_KEY
] = max(result
[NODE_MAX_DEPTH_KEY
], depth
)
185 if not seen_symbol_with_path
:
186 logging
.warning('Symbols lack paths. Data will not be structured.')
188 # The (no path) bucket can be extremely large if we failed to get
189 # path information. Split it into subgroups if needed.
190 SplitNoPathBucket(result
)
192 largest_list_len
= MakeChildrenDictsIntoLists(result
)
194 if largest_list_len
> BIG_BUCKET_LIMIT
:
195 logging
.warning('There are sections with %d nodes. '
196 'Results might be unusable.' % largest_list_len
)
200 def DumpCompactTree(symbols
, symbol_path_origin_dir
, outfile
):
201 tree_root
= MakeCompactTree(symbols
, symbol_path_origin_dir
)
202 with
open(outfile
, 'w') as out
:
203 out
.write('var tree_data=')
204 # Use separators without whitespace to get a smaller file.
205 json
.dump(tree_root
, out
, separators
=(',', ':'))
206 print('Writing %d bytes json' % os
.path
.getsize(outfile
))
209 def MakeSourceMap(symbols
):
211 for _sym
, _symbol_type
, size
, path
, _address
in symbols
:
214 key
= os
.path
.normpath(path
)
217 if key
not in sources
:
218 sources
[key
] = {'path': path
, 'symbol_count': 0, 'size': 0}
219 record
= sources
[key
]
220 record
['size'] += size
221 record
['symbol_count'] += 1
225 # Regex for parsing "nm" output. A sample line looks like this:
226 # 0167b39c 00000018 t ACCESS_DESCRIPTION_free /path/file.c:95
228 # The fields are: address, size, type, name, source location
229 # Regular expression explained ( see also: https://xkcd.com/208 ):
230 # ([0-9a-f]{8,}+) The address
231 # [\s]+ Whitespace separator
232 # ([0-9a-f]{8,}+) The size. From here on out it's all optional.
233 # [\s]+ Whitespace separator
234 # (\S?) The symbol type, which is any non-whitespace char
235 # [\s*] Whitespace separator
236 # ([^\t]*) Symbol name, any non-tab character (spaces ok!)
237 # [\t]? Tab separator
238 # (.*) The location (filename[:linennum|?][ (discriminator n)]
239 sNmPattern
= re
.compile(
240 r
'([0-9a-f]{8,})[\s]+([0-9a-f]{8,})[\s]*(\S?)[\s*]([^\t]*)[\t]?(.*)')
247 self
.time_last_output
= time
.time()
248 self
.count_last_output
= 0
249 self
.disambiguations
= 0
250 self
.was_ambiguous
= 0
253 def RunElfSymbolizer(outfile
, library
, addr2line_binary
, nm_binary
, jobs
,
254 disambiguate
, src_path
):
255 nm_output
= RunNm(library
, nm_binary
)
256 nm_output_lines
= nm_output
.splitlines()
257 nm_output_lines_len
= len(nm_output_lines
)
259 progress
= Progress()
260 def map_address_symbol(symbol
, addr
):
262 if addr
in address_symbol
:
263 # 'Collision between %s and %s.' % (str(symbol.name),
264 # str(address_symbol[addr].name))
265 progress
.collisions
+= 1
267 if symbol
.disambiguated
:
268 progress
.disambiguations
+= 1
269 if symbol
.was_ambiguous
:
270 progress
.was_ambiguous
+= 1
272 address_symbol
[addr
] = symbol
276 def progress_output():
278 if progress
.count
% progress_chunk
== 0:
279 time_now
= time
.time()
280 time_spent
= time_now
- progress
.time_last_output
282 # Only output at most once per second.
283 progress
.time_last_output
= time_now
284 chunk_size
= progress
.count
- progress
.count_last_output
285 progress
.count_last_output
= progress
.count
287 speed
= chunk_size
/ time_spent
290 progress_percent
= (100.0 * (progress
.count
+ progress
.skip_count
) /
292 disambiguation_percent
= 0
293 if progress
.disambiguations
!= 0:
294 disambiguation_percent
= (100.0 * progress
.disambiguations
/
295 progress
.was_ambiguous
)
297 sys
.stdout
.write('\r%.1f%%: Looked up %d symbols (%d collisions, '
298 '%d disambiguations where %.1f%% succeeded)'
299 ' - %.1f lookups/s.' %
300 (progress_percent
, progress
.count
, progress
.collisions
,
301 progress
.disambiguations
, disambiguation_percent
, speed
))
303 # In case disambiguation was disabled, we remove the source path (which upon
304 # being set signals the symbolizer to enable disambiguation)
307 symbolizer
= elf_symbolizer
.ELFSymbolizer(library
, addr2line_binary
,
309 max_concurrent_jobs
=jobs
,
310 source_root_path
=src_path
)
311 user_interrupted
= False
313 for line
in nm_output_lines
:
314 match
= sNmPattern
.match(line
)
316 location
= match
.group(5)
318 addr
= int(match
.group(1), 16)
319 size
= int(match
.group(2), 16)
320 if addr
in address_symbol
: # Already looked up, shortcut
322 map_address_symbol(address_symbol
[addr
], addr
)
325 # Save time by not looking up empty symbols (do they even exist?)
326 print('Empty symbol: ' + line
)
328 symbolizer
.SymbolizeAsync(addr
, addr
)
331 progress
.skip_count
+= 1
332 except KeyboardInterrupt:
333 user_interrupted
= True
334 print('Interrupting - killing subprocesses. Please wait.')
338 except KeyboardInterrupt:
339 # Don't want to abort here since we will be finished in a few seconds.
340 user_interrupted
= True
341 print('Patience you must have my young padawan.')
346 print('Skipping the rest of the file mapping. '
347 'Output will not be fully classified.')
349 symbol_path_origin_dir
= os
.path
.dirname(os
.path
.abspath(library
))
351 with
open(outfile
, 'w') as out
:
352 for line
in nm_output_lines
:
353 match
= sNmPattern
.match(line
)
355 location
= match
.group(5)
357 addr
= int(match
.group(1), 16)
358 symbol
= address_symbol
.get(addr
)
359 if symbol
is not None:
361 if symbol
.source_path
is not None:
362 path
= os
.path
.abspath(os
.path
.join(symbol_path_origin_dir
,
365 if symbol
.source_line
is not None:
366 line_number
= symbol
.source_line
367 out
.write('%s\t%s:%d\n' % (line
, path
, line_number
))
370 out
.write('%s\n' % line
)
372 print('%d symbols in the results.' % len(address_symbol
))
375 def RunNm(binary
, nm_binary
):
376 cmd
= [nm_binary
, '-C', '--print-size', '--size-sort', '--reverse-sort',
378 nm_process
= subprocess
.Popen(cmd
,
379 stdout
=subprocess
.PIPE
,
380 stderr
=subprocess
.PIPE
)
381 (process_output
, err_output
) = nm_process
.communicate()
383 if nm_process
.returncode
!= 0:
385 raise Exception, err_output
387 raise Exception, process_output
389 return process_output
392 def GetNmSymbols(nm_infile
, outfile
, library
, jobs
, verbose
,
393 addr2line_binary
, nm_binary
, disambiguate
, src_path
):
394 if nm_infile
is None:
396 outfile
= tempfile
.NamedTemporaryFile(delete
=False).name
399 print 'Running parallel addr2line, dumping symbols to ' + outfile
400 RunElfSymbolizer(outfile
, library
, addr2line_binary
, nm_binary
, jobs
,
401 disambiguate
, src_path
)
406 print 'Using nm input from ' + nm_infile
407 with
file(nm_infile
, 'r') as infile
:
408 return list(binary_size_utils
.ParseNm(infile
))
411 PAK_RESOURCE_ID_TO_STRING
= { "inited": False }
413 def LoadPakIdsFromResourceFile(filename
):
414 """Given a file name, it loads everything that looks like a resource id
415 into PAK_RESOURCE_ID_TO_STRING."""
416 with
open(filename
) as resource_header
:
417 for line
in resource_header
:
418 if line
.startswith("#define "):
419 line_data
= line
.split()
420 if len(line_data
) == 3:
422 resource_number
= int(line_data
[2])
423 resource_name
= line_data
[1]
424 PAK_RESOURCE_ID_TO_STRING
[resource_number
] = resource_name
428 def GetReadablePakResourceName(pak_file
, resource_id
):
429 """Pak resources have a numeric identifier. It is not helpful when
430 trying to locate where footprint is generated. This does its best to
431 map the number to a usable string."""
432 if not PAK_RESOURCE_ID_TO_STRING
['inited']:
433 # Try to find resource header files generated by grit when
434 # building the pak file. We'll look for files named *resources.h"
435 # and lines of the type:
436 # #define MY_RESOURCE_JS 1234
437 PAK_RESOURCE_ID_TO_STRING
['inited'] = True
438 gen_dir
= os
.path
.join(os
.path
.dirname(pak_file
), 'gen')
439 if os
.path
.isdir(gen_dir
):
440 for dirname
, _dirs
, files
in os
.walk(gen_dir
):
441 for filename
in files
:
442 if filename
.endswith('resources.h'):
443 LoadPakIdsFromResourceFile(os
.path
.join(dirname
, filename
))
444 return PAK_RESOURCE_ID_TO_STRING
.get(resource_id
,
445 'Pak Resource %d' % resource_id
)
447 def AddPakData(symbols
, pak_file
):
448 """Adds pseudo-symbols from a pak file."""
449 pak_file
= os
.path
.abspath(pak_file
)
450 with
open(pak_file
, 'rb') as pak
:
454 HEADER_LENGTH
= 2 * 4 + 1 # Two uint32s. (file version, number of entries)
455 # and one uint8 (encoding of text resources)
456 INDEX_ENTRY_SIZE
= 2 + 4 # Each entry is a uint16 and a uint32.
457 version
, num_entries
, _encoding
= struct
.unpack('<IIB', data
[:HEADER_LENGTH
])
458 assert version
== PAK_FILE_VERSION
, ('Unsupported pak file '
459 'version (%d) in %s. Only '
460 'support version %d' %
461 (version
, pak_file
, PAK_FILE_VERSION
))
463 # Read the index and data.
464 data
= data
[HEADER_LENGTH
:]
465 for _
in range(num_entries
):
466 resource_id
, offset
= struct
.unpack('<HI', data
[:INDEX_ENTRY_SIZE
])
467 data
= data
[INDEX_ENTRY_SIZE
:]
468 _next_id
, next_offset
= struct
.unpack('<HI', data
[:INDEX_ENTRY_SIZE
])
469 resource_size
= next_offset
- offset
471 symbol_name
= GetReadablePakResourceName(pak_file
, resource_id
)
472 symbol_path
= pak_file
473 symbol_type
= 'd' # Data. Approximation.
474 symbol_size
= resource_size
475 symbols
.append((symbol_name
, symbol_type
, symbol_size
, symbol_path
))
477 def _find_in_system_path(binary
):
478 """Locate the full path to binary in the system path or return None
480 system_path
= os
.environ
["PATH"].split(os
.pathsep
)
481 for path
in system_path
:
482 binary_path
= os
.path
.join(path
, binary
)
483 if os
.path
.isfile(binary_path
):
487 def CheckDebugFormatSupport(library
, addr2line_binary
):
488 """Kills the program if debug data is in an unsupported format.
490 There are two common versions of the DWARF debug formats and
491 since we are right now transitioning from DWARF2 to newer formats,
492 it's possible to have a mix of tools that are not compatible. Detect
493 that and abort rather than produce meaningless output."""
494 tool_output
= subprocess
.check_output([addr2line_binary
, '--version'])
495 version_re
= re
.compile(r
'^GNU [^ ]+ .* (\d+).(\d+).*?$', re
.M
)
496 parsed_output
= version_re
.match(tool_output
)
497 major
= int(parsed_output
.group(1))
498 minor
= int(parsed_output
.group(2))
499 supports_dwarf4
= major
> 2 or major
== 2 and minor
> 22
504 print('Checking version of debug information in %s.' % library
)
505 debug_info
= subprocess
.check_output(['readelf', '--debug-dump=info',
506 '--dwarf-depth=1', library
])
507 dwarf_version_re
= re
.compile(r
'^\s+Version:\s+(\d+)$', re
.M
)
508 parsed_dwarf_format_output
= dwarf_version_re
.search(debug_info
)
509 version
= int(parsed_dwarf_format_output
.group(1))
511 print('The supplied tools only support DWARF2 debug data but the binary\n' +
512 'uses DWARF%d. Update the tools or compile the binary\n' % version
+
518 usage
= """%prog [options]
520 Runs a spatial analysis on a given library, looking up the source locations
521 of its symbols and calculating how much space each directory, source file,
522 and so on is taking. The result is a report that can be used to pinpoint
523 sources of large portions of the binary, etceteras.
525 Under normal circumstances, you only need to pass two arguments, thusly:
527 %prog --library /path/to/library --destdir /path/to/output
529 In this mode, the program will dump the symbols from the specified library
530 and map those symbols back to source locations, producing a web-based
531 report in the specified output directory.
533 Other options are available via '--help'.
535 parser
= optparse
.OptionParser(usage
=usage
)
536 parser
.add_option('--nm-in', metavar
='PATH',
537 help='if specified, use nm input from <path> instead of '
538 'generating it. Note that source locations should be '
539 'present in the file; i.e., no addr2line symbol lookups '
540 'will be performed when this option is specified. '
541 'Mutually exclusive with --library.')
542 parser
.add_option('--destdir', metavar
='PATH',
543 help='write output to the specified directory. An HTML '
544 'report is generated here along with supporting files; '
545 'any existing report will be overwritten.')
546 parser
.add_option('--library', metavar
='PATH',
547 help='if specified, process symbols in the library at '
548 'the specified path. Mutually exclusive with --nm-in.')
549 parser
.add_option('--pak', metavar
='PATH',
550 help='if specified, includes the contents of the '
551 'specified *.pak file in the output.')
552 parser
.add_option('--nm-binary',
553 help='use the specified nm binary to analyze library. '
554 'This is to be used when the nm in the path is not for '
555 'the right architecture or of the right version.')
556 parser
.add_option('--addr2line-binary',
557 help='use the specified addr2line binary to analyze '
558 'library. This is to be used when the addr2line in '
559 'the path is not for the right architecture or '
560 'of the right version.')
561 parser
.add_option('--jobs', type='int',
562 help='number of jobs to use for the parallel '
563 'addr2line processing pool; defaults to 1. More '
564 'jobs greatly improve throughput but eat RAM like '
565 'popcorn, and take several gigabytes each. Start low '
566 'and ramp this number up until your machine begins to '
567 'struggle with RAM. '
568 'This argument is only valid when using --library.')
569 parser
.add_option('-v', '--verbose', dest
='verbose', action
='store_true',
570 help='be verbose, printing lots of status information.')
571 parser
.add_option('--nm-out', metavar
='PATH',
572 help='(deprecated) No-op. nm.out is stored in --destdir.')
573 parser
.add_option('--no-nm-out', action
='store_true',
574 help='do not keep the nm output file. This file is useful '
575 'if you want to see the fully processed nm output after '
576 'the symbols have been mapped to source locations, or if '
577 'you plan to run explain_binary_size_delta.py. By default '
578 'the file \'nm.out\' is placed alongside the generated '
579 'report. The nm.out file is only created when using '
581 parser
.add_option('--disable-disambiguation', action
='store_true',
582 help='disables the disambiguation process altogether,'
583 ' NOTE: this may, depending on your toolchain, produce'
584 ' output with some symbols at the top layer if addr2line'
585 ' could not get the entire source path.')
586 parser
.add_option('--source-path', default
='./',
587 help='the path to the source code of the output binary, '
588 'default set to current directory. Used in the'
589 ' disambiguation process.')
590 opts
, _args
= parser
.parse_args()
592 if ((not opts
.library
) and (not opts
.nm_in
)) or (opts
.library
and opts
.nm_in
):
593 parser
.error('exactly one of --library or --nm-in is required')
595 print >> sys
.stderr
, ('WARNING: --nm-out is deprecated and has no effect.')
598 print >> sys
.stderr
, ('WARNING: --jobs has no effect '
599 'when used with --nm-in')
601 parser
.error('--destdir is a required argument')
603 # Use the number of processors but cap between 2 and 4 since raw
604 # CPU power isn't the limiting factor. It's I/O limited, memory
605 # bus limited and available-memory-limited. Too many processes and
606 # the computer will run out of memory and it will be slow.
607 opts
.jobs
= max(2, min(4, str(multiprocessing
.cpu_count())))
609 if opts
.addr2line_binary
:
610 assert os
.path
.isfile(opts
.addr2line_binary
)
611 addr2line_binary
= opts
.addr2line_binary
613 addr2line_binary
= _find_in_system_path('addr2line')
614 assert addr2line_binary
, 'Unable to find addr2line in the path. '\
615 'Use --addr2line-binary to specify location.'
618 assert os
.path
.isfile(opts
.nm_binary
)
619 nm_binary
= opts
.nm_binary
621 nm_binary
= _find_in_system_path('nm')
622 assert nm_binary
, 'Unable to find nm in the path. Use --nm-binary '\
623 'to specify location.'
626 assert os
.path
.isfile(opts
.pak
), 'Could not find ' % opts
.pak
628 print('addr2line: %s' % addr2line_binary
)
629 print('nm: %s' % nm_binary
)
632 CheckDebugFormatSupport(opts
.library
, addr2line_binary
)
634 # Prepare output directory and report guts
635 if not os
.path
.exists(opts
.destdir
):
636 os
.makedirs(opts
.destdir
, 0755)
637 nm_out
= os
.path
.join(opts
.destdir
, 'nm.out')
641 # Copy report boilerplate into output directory. This also proves that the
642 # output directory is safe for writing, so there should be no problems writing
643 # the nm.out file later.
644 data_js_file_name
= os
.path
.join(opts
.destdir
, 'data.js')
645 d3_out
= os
.path
.join(opts
.destdir
, 'd3')
646 if not os
.path
.exists(d3_out
):
647 os
.makedirs(d3_out
, 0755)
648 d3_src
= os
.path
.join(os
.path
.dirname(__file__
),
651 'third_party', 'd3', 'src')
652 template_src
= os
.path
.join(os
.path
.dirname(__file__
),
654 shutil
.copy(os
.path
.join(d3_src
, 'LICENSE'), d3_out
)
655 shutil
.copy(os
.path
.join(d3_src
, 'd3.js'), d3_out
)
656 shutil
.copy(os
.path
.join(template_src
, 'index.html'), opts
.destdir
)
657 shutil
.copy(os
.path
.join(template_src
, 'D3SymbolTreeMap.js'), opts
.destdir
)
659 # Run nm and/or addr2line to gather the data
660 symbols
= GetNmSymbols(opts
.nm_in
, nm_out
, opts
.library
,
661 opts
.jobs
, opts
.verbose
is True,
662 addr2line_binary
, nm_binary
,
663 opts
.disable_disambiguation
is None,
668 AddPakData(symbols
, opts
.pak
)
670 symbol_path_origin_dir
= os
.path
.dirname(os
.path
.abspath(opts
.library
))
672 # Just a guess. Hopefully all paths in the input file are absolute.
673 symbol_path_origin_dir
= os
.path
.abspath(os
.getcwd())
674 # Dump JSON for the HTML report.
675 DumpCompactTree(symbols
, symbol_path_origin_dir
, data_js_file_name
)
676 print 'Report saved to ' + opts
.destdir
+ '/index.html'
678 if __name__
== '__main__':