1 /* Profiling of shared libraries.
2 Copyright (C) 1997-2022 Free Software Foundation, Inc.
3 This file is part of the GNU C Library.
4 Based on the BSD mcount implementation.
6 The GNU C Library is free software; you can redistribute it and/or
7 modify it under the terms of the GNU Lesser General Public
8 License as published by the Free Software Foundation; either
9 version 2.1 of the License, or (at your option) any later version.
11 The GNU C Library is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 Lesser General Public License for more details.
16 You should have received a copy of the GNU Lesser General Public
17 License along with the GNU C Library; if not, see
18 <https://www.gnu.org/licenses/>. */
32 #include <sys/gmon_out.h>
34 #include <sys/param.h>
37 #include <not-cancel.h>
39 /* The LD_PROFILE feature has to be implemented different to the
40 normal profiling using the gmon/ functions. The problem is that an
41 arbitrary amount of processes simulataneously can be run using
42 profiling and all write the results in the same file. To provide
43 this mechanism one could implement a complicated mechanism to merge
44 the content of two profiling runs or one could extend the file
45 format to allow more than one data set. For the second solution we
46 would have the problem that the file can grow in size beyond any
47 limit and both solutions have the problem that the concurrency of
48 writing the results is a big problem.
50 Another much simpler method is to use mmap to map the same file in
51 all using programs and modify the data in the mmap'ed area and so
52 also automatically on the disk. Using the MAP_SHARED option of
53 mmap(2) this can be done without big problems in more than one
56 This approach is very different from the normal profiling. We have
57 to use the profiling data in exactly the way they are expected to
58 be written to disk. But the normal format used by gprof is not usable
59 to do this. It is optimized for size. It writes the tags as single
60 bytes but this means that the following 32/64 bit values are
63 Therefore we use a new format. This will look like this
65 0 1 2 3 <- byte is 32 bit word
67 0004 *version* <- GMON_SHOBJ_VERSION
72 0014 *tag* <- GMON_TAG_TIME_HIST
74 ?? ?? ?? ?? <- 32/64 bit LowPC
76 ?? ?? ?? ?? <- 32/64 bit HighPC
85 0030+2*A ?? ?? ?? ?? <- Count data
87 0030+2*A+K ?? ?? ?? ??
89 0030+2*A+K *tag* <- GMON_TAG_CG_ARC
91 0038+2*A+K ?? ?? ?? ??
92 ?? ?? ?? ?? <- FromPC#1
93 0038+3*A+K ?? ?? ?? ??
95 0038+4*A+K ?? ?? ?? ?? <- Count#1
97 0038+(2*(CN-1)+2)*A+(CN-1)*4+K ?? ?? ?? ??
98 ?? ?? ?? ?? <- FromPC#CGN
99 0038+(2*(CN-1)+3)*A+(CN-1)*4+K ?? ?? ?? ??
100 ?? ?? ?? ?? <- ToPC#CGN
101 0038+(2*CN+2)*A+(CN-1)*4+K ?? ?? ?? ?? <- Count#CGN
103 We put (for now?) no basic block information in the file since this would
104 introduce rase conditions among all the processes who want to write them.
106 `K' is the number of count entries which is computed as
108 textsize / HISTFRACTION
110 `CG' in the above table is the number of call graph arcs. Normally,
111 the table is sparse and the profiling code writes out only the those
112 entries which are really used in the program run. But since we must
113 not extend this table (the profiling file) we'll keep them all here.
114 So CN can be executed in advance as
116 MINARCS <= textsize*(ARCDENSITY/100) <= MAXARCS
118 Now the remaining question is: how to build the data structures we can
119 work with from this data. We need the from set and must associate the
120 froms with all the associated tos. We will do this by constructing this
121 data structures at the program start. To do this we'll simply visit all
122 entries in the call graph table and add it to the appropriate list. */
124 extern int __profile_frequency (void);
125 libc_hidden_proto (__profile_frequency
)
127 /* We define a special type to address the elements of the arc table.
128 This is basically the `gmon_cg_arc_record' format but it includes
129 the room for the tag and it uses real types. */
130 struct here_cg_arc_record
134 /* The count field is atomically incremented in _dl_mcount, which
135 requires it to be properly aligned for its type, and for this
136 alignment to be visible to the compiler. The amount of data
137 before an array of this structure is calculated as
138 expected_size in _dl_start_profile. Everything in that
139 calculation is a multiple of 4 bytes (in the case of
140 kcountsize, because it is derived from a subtraction of
141 page-aligned values, and the corresponding calculation in
142 __monstartup also ensures it is at least a multiple of the size
143 of u_long), so all copies of this field do in fact have the
144 appropriate alignment. */
145 uint32_t count
__attribute__ ((aligned (__alignof__ (uint32_t))));
146 } __attribute__ ((packed
));
148 static struct here_cg_arc_record
*data
;
150 /* Nonzero if profiling is under way. */
153 /* This is the number of entry which have been incorporated in the toset. */
154 static uint32_t narcs
;
155 /* This is a pointer to the object representing the number of entries
156 currently in the mmaped file. At no point of time this has to be the
157 same as NARCS. If it is equal all entries from the file are in our
159 static volatile uint32_t *narcsp
;
162 struct here_fromstruct
164 struct here_cg_arc_record
volatile *here
;
168 static volatile uint16_t *tos
;
170 static struct here_fromstruct
*froms
;
171 static uint32_t fromlimit
;
172 static volatile uint32_t fromidx
;
174 static uintptr_t lowpc
;
175 static size_t textsize
;
176 static unsigned int log_hashfraction
;
180 /* Set up profiling data to profile object desribed by MAP. The output
181 file is found (or created) in OUTPUT_DIR. */
183 _dl_start_profile (void)
187 struct __stat64_t64 st
;
188 const ElfW(Phdr
) *ph
;
189 ElfW(Addr
) mapstart
= ~((ElfW(Addr
)) 0);
190 ElfW(Addr
) mapend
= 0;
198 struct gmon_hdr
*addr
= NULL
;
200 /* See profil(2) where this is described. */
202 #define SCALE_1_TO_1 0x10000L
203 const char *errstr
= NULL
;
205 /* Compute the size of the sections which contain program code. */
206 for (ph
= GL(dl_profile_map
)->l_phdr
;
207 ph
< &GL(dl_profile_map
)->l_phdr
[GL(dl_profile_map
)->l_phnum
]; ++ph
)
208 if (ph
->p_type
== PT_LOAD
&& (ph
->p_flags
& PF_X
))
210 ElfW(Addr
) start
= (ph
->p_vaddr
& ~(GLRO(dl_pagesize
) - 1));
211 ElfW(Addr
) end
= ((ph
->p_vaddr
+ ph
->p_memsz
+ GLRO(dl_pagesize
) - 1)
212 & ~(GLRO(dl_pagesize
) - 1));
214 if (start
< mapstart
)
220 /* Now we can compute the size of the profiling data. This is done
221 with the same formulars as in `monstartup' (see gmon.c). */
223 lowpc
= ROUNDDOWN (mapstart
+ GL(dl_profile_map
)->l_addr
,
224 HISTFRACTION
* sizeof (HISTCOUNTER
));
225 highpc
= ROUNDUP (mapend
+ GL(dl_profile_map
)->l_addr
,
226 HISTFRACTION
* sizeof (HISTCOUNTER
));
227 textsize
= highpc
- lowpc
;
228 kcountsize
= textsize
/ HISTFRACTION
;
229 if ((HASHFRACTION
& (HASHFRACTION
- 1)) == 0)
231 /* If HASHFRACTION is a power of two, mcount can use shifting
232 instead of integer division. Precompute shift amount.
234 This is a constant but the compiler cannot compile the
235 expression away since the __ffs implementation is not known
236 to the compiler. Help the compiler by precomputing the
238 assert (HASHFRACTION
== 2);
240 if (sizeof (*froms
) == 8)
241 log_hashfraction
= 4;
242 else if (sizeof (*froms
) == 16)
243 log_hashfraction
= 5;
245 log_hashfraction
= __ffs (HASHFRACTION
* sizeof (*froms
)) - 1;
248 log_hashfraction
= -1;
249 tossize
= textsize
/ HASHFRACTION
;
250 fromlimit
= textsize
* ARCDENSITY
/ 100;
251 if (fromlimit
< MINARCS
)
253 if (fromlimit
> MAXARCS
)
255 fromssize
= fromlimit
* sizeof (struct here_fromstruct
);
257 expected_size
= (sizeof (struct gmon_hdr
)
258 + 4 + sizeof (struct gmon_hist_hdr
) + kcountsize
259 + 4 + 4 + fromssize
* sizeof (struct here_cg_arc_record
));
261 /* Create the gmon_hdr we expect or write. */
268 if (sizeof (gmon_hdr
) != sizeof (struct gmon_hdr
)
269 || (offsetof (struct real_gmon_hdr
, cookie
)
270 != offsetof (struct gmon_hdr
, cookie
))
271 || (offsetof (struct real_gmon_hdr
, version
)
272 != offsetof (struct gmon_hdr
, version
)))
275 memcpy (&gmon_hdr
.cookie
[0], GMON_MAGIC
, sizeof (gmon_hdr
.cookie
));
276 gmon_hdr
.version
= GMON_SHOBJ_VERSION
;
277 memset (gmon_hdr
.spare
, '\0', sizeof (gmon_hdr
.spare
));
279 /* Create the hist_hdr we expect or write. */
280 struct real_gmon_hist_hdr
289 if (sizeof (hist_hdr
) != sizeof (struct gmon_hist_hdr
)
290 || (offsetof (struct real_gmon_hist_hdr
, low_pc
)
291 != offsetof (struct gmon_hist_hdr
, low_pc
))
292 || (offsetof (struct real_gmon_hist_hdr
, high_pc
)
293 != offsetof (struct gmon_hist_hdr
, high_pc
))
294 || (offsetof (struct real_gmon_hist_hdr
, hist_size
)
295 != offsetof (struct gmon_hist_hdr
, hist_size
))
296 || (offsetof (struct real_gmon_hist_hdr
, prof_rate
)
297 != offsetof (struct gmon_hist_hdr
, prof_rate
))
298 || (offsetof (struct real_gmon_hist_hdr
, dimen
)
299 != offsetof (struct gmon_hist_hdr
, dimen
))
300 || (offsetof (struct real_gmon_hist_hdr
, dimen_abbrev
)
301 != offsetof (struct gmon_hist_hdr
, dimen_abbrev
)))
304 hist_hdr
.low_pc
= (char *) mapstart
;
305 hist_hdr
.high_pc
= (char *) mapend
;
306 hist_hdr
.hist_size
= kcountsize
/ sizeof (HISTCOUNTER
);
307 hist_hdr
.prof_rate
= __profile_frequency ();
308 if (sizeof (hist_hdr
.dimen
) >= sizeof ("seconds"))
310 memcpy (hist_hdr
.dimen
, "seconds", sizeof ("seconds"));
311 memset (hist_hdr
.dimen
+ sizeof ("seconds"), '\0',
312 sizeof (hist_hdr
.dimen
) - sizeof ("seconds"));
315 strncpy (hist_hdr
.dimen
, "seconds", sizeof (hist_hdr
.dimen
));
316 hist_hdr
.dimen_abbrev
= 's';
318 /* First determine the output name. We write in the directory
319 OUTPUT_DIR and the name is composed from the shared objects
320 soname (or the file name) and the ending ".profile". */
321 filename
= (char *) alloca (strlen (GLRO(dl_profile_output
)) + 1
322 + strlen (GLRO(dl_profile
)) + sizeof ".profile");
323 cp
= __stpcpy (filename
, GLRO(dl_profile_output
));
325 __stpcpy (__stpcpy (cp
, GLRO(dl_profile
)), ".profile");
327 fd
= __open64_nocancel (filename
, O_RDWR
|O_CREAT
|O_NOFOLLOW
, DEFFILEMODE
);
333 /* We cannot write the profiling data so don't do anything. */
334 errstr
= "%s: cannot open file: %s\n";
338 __close_nocancel (fd
);
339 _dl_error_printf (errstr
, filename
,
340 __strerror_r (errnum
, buf
, sizeof buf
));
344 if (__fstat64_time64 (fd
, &st
) < 0 || !S_ISREG (st
.st_mode
))
346 /* Not stat'able or not a regular file => don't use it. */
347 errstr
= "%s: cannot stat file: %s\n";
351 /* Test the size. If it does not match what we expect from the size
352 values in the map MAP we don't use it and warn the user. */
355 /* We have to create the file. */
356 char buf
[GLRO(dl_pagesize
)];
358 memset (buf
, '\0', GLRO(dl_pagesize
));
360 if (__lseek (fd
, expected_size
& ~(GLRO(dl_pagesize
) - 1), SEEK_SET
) == -1)
363 errstr
= "%s: cannot create file: %s\n";
367 if (TEMP_FAILURE_RETRY
368 (__write_nocancel (fd
, buf
, (expected_size
& (GLRO(dl_pagesize
) - 1))))
372 else if (st
.st_size
!= expected_size
)
374 __close_nocancel (fd
);
378 __munmap ((void *) addr
, expected_size
);
380 _dl_error_printf ("%s: file is no correct profile data file for `%s'\n",
381 filename
, GLRO(dl_profile
));
385 addr
= (struct gmon_hdr
*) __mmap (NULL
, expected_size
, PROT_READ
|PROT_WRITE
,
386 MAP_SHARED
|MAP_FILE
, fd
, 0);
387 if (addr
== (struct gmon_hdr
*) MAP_FAILED
)
389 errstr
= "%s: cannot map file: %s\n";
393 /* We don't need the file descriptor anymore. */
394 __close_nocancel (fd
);
396 /* Pointer to data after the header. */
397 hist
= (char *) (addr
+ 1);
398 kcount
= (uint16_t *) ((char *) hist
+ sizeof (uint32_t)
399 + sizeof (struct gmon_hist_hdr
));
401 /* Compute pointer to array of the arc information. */
402 narcsp
= (uint32_t *) ((char *) kcount
+ kcountsize
+ sizeof (uint32_t));
403 data
= (struct here_cg_arc_record
*) ((char *) narcsp
+ sizeof (uint32_t));
407 /* Create the signature. */
408 memcpy (addr
, &gmon_hdr
, sizeof (struct gmon_hdr
));
410 *(uint32_t *) hist
= GMON_TAG_TIME_HIST
;
411 memcpy (hist
+ sizeof (uint32_t), &hist_hdr
,
412 sizeof (struct gmon_hist_hdr
));
414 narcsp
[-1] = GMON_TAG_CG_ARC
;
418 /* Test the signature in the file. */
419 if (memcmp (addr
, &gmon_hdr
, sizeof (struct gmon_hdr
)) != 0
420 || *(uint32_t *) hist
!= GMON_TAG_TIME_HIST
421 || memcmp (hist
+ sizeof (uint32_t), &hist_hdr
,
422 sizeof (struct gmon_hist_hdr
)) != 0
423 || narcsp
[-1] != GMON_TAG_CG_ARC
)
427 /* Allocate memory for the froms data and the pointer to the tos records. */
428 tos
= (uint16_t *) calloc (tossize
+ fromssize
, 1);
431 __munmap ((void *) addr
, expected_size
);
432 _dl_fatal_printf ("Out of memory while initializing profiler\n");
436 froms
= (struct here_fromstruct
*) ((char *) tos
+ tossize
);
439 /* Now we have to process all the arc count entries. BTW: it is
440 not critical whether the *NARCSP value changes meanwhile. Before
441 we enter a new entry in to toset we will check that everything is
442 available in TOS. This happens in _dl_mcount.
444 Loading the entries in reverse order should help to get the most
445 frequently used entries at the front of the list. */
446 for (idx
= narcs
= MIN (*narcsp
, fromlimit
); idx
> 0; )
451 to_index
= (data
[idx
].self_pc
/ (HASHFRACTION
* sizeof (*tos
)));
452 newfromidx
= fromidx
++;
453 froms
[newfromidx
].here
= &data
[idx
];
454 froms
[newfromidx
].link
= tos
[to_index
];
455 tos
[to_index
] = newfromidx
;
458 /* Setup counting data. */
459 if (kcountsize
< highpc
- lowpc
)
462 s_scale
= ((double) kcountsize
/ (highpc
- lowpc
)) * SCALE_1_TO_1
;
464 size_t range
= highpc
- lowpc
;
465 size_t quot
= range
/ kcountsize
;
467 if (quot
>= SCALE_1_TO_1
)
469 else if (quot
>= SCALE_1_TO_1
/ 256)
470 s_scale
= SCALE_1_TO_1
/ quot
;
471 else if (range
> ULONG_MAX
/ 256)
472 s_scale
= (SCALE_1_TO_1
* 256) / (range
/ (kcountsize
/ 256));
474 s_scale
= (SCALE_1_TO_1
* 256) / ((range
* 256) / kcountsize
);
478 s_scale
= SCALE_1_TO_1
;
480 /* Start the profiler. */
481 __profil ((void *) kcount
, kcountsize
, lowpc
, s_scale
);
483 /* Turn on profiling. */
489 _dl_mcount (ElfW(Addr
) frompc
, ElfW(Addr
) selfpc
)
491 volatile uint16_t *topcindex
;
493 struct here_fromstruct
*fromp
;
498 /* Compute relative addresses. The shared object can be loaded at
499 any address. The value of frompc could be anything. We cannot
500 restrict it in any way, just set to a fixed value (0) in case it
501 is outside the allowed range. These calls show up as calls from
502 <external> in the gprof output. */
504 if (frompc
>= textsize
)
507 if (selfpc
>= textsize
)
510 /* Getting here we now have to find out whether the location was
511 already used. If yes we are lucky and only have to increment a
512 counter (this also has to be atomic). If the entry is new things
513 are getting complicated... */
515 /* Avoid integer divide if possible. */
516 if ((HASHFRACTION
& (HASHFRACTION
- 1)) == 0)
517 i
= selfpc
>> log_hashfraction
;
519 i
= selfpc
/ (HASHFRACTION
* sizeof (*tos
));
522 fromindex
= *topcindex
;
525 goto check_new_or_add
;
527 fromp
= &froms
[fromindex
];
529 /* We have to look through the chain of arcs whether there is already
530 an entry for our arc. */
531 while (fromp
->here
->from_pc
!= frompc
)
533 if (fromp
->link
!= 0)
535 fromp
= &froms
[fromp
->link
];
536 while (fromp
->link
!= 0 && fromp
->here
->from_pc
!= frompc
);
538 if (fromp
->here
->from_pc
!= frompc
)
540 topcindex
= &fromp
->link
;
543 /* Our entry is not among the entries we read so far from the
544 data file. Now see whether we have to update the list. */
545 while (narcs
!= *narcsp
&& narcs
< fromlimit
)
549 to_index
= (data
[narcs
].self_pc
550 / (HASHFRACTION
* sizeof (*tos
)));
551 newfromidx
= atomic_fetch_add_relaxed (&fromidx
, 1) + 1;
552 froms
[newfromidx
].here
= &data
[narcs
];
553 froms
[newfromidx
].link
= tos
[to_index
];
554 tos
[to_index
] = newfromidx
;
555 catomic_increment (&narcs
);
558 /* If we still have no entry stop searching and insert. */
561 unsigned int newarc
= atomic_fetch_add_relaxed (narcsp
, 1);
563 /* In rare cases it could happen that all entries in FROMS are
564 occupied. So we cannot count this anymore. */
565 if (newarc
>= fromlimit
)
568 *topcindex
= atomic_fetch_add_relaxed (&fromidx
, 1) + 1;
569 fromp
= &froms
[*topcindex
];
571 fromp
->here
= &data
[newarc
];
572 data
[newarc
].from_pc
= frompc
;
573 data
[newarc
].self_pc
= selfpc
;
574 data
[newarc
].count
= 0;
576 atomic_fetch_add_relaxed (&narcs
, 1);
581 fromp
= &froms
[*topcindex
];
588 /* Increment the counter. */
589 atomic_fetch_add_relaxed (&fromp
->here
->count
, 1);
594 rtld_hidden_def (_dl_mcount
)