1 /* Profile counter container type.
2 Copyright (C) 2017-2023 Free Software Foundation, Inc.
3 Contributed by Jan Hubicka
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 #ifndef GCC_PROFILE_COUNT_H
22 #define GCC_PROFILE_COUNT_H
28 /* Quality of the profile count. Because gengtype does not support enums
29 inside of classes, this is in global namespace. */
30 enum profile_quality
{
31 /* Uninitialized value. */
32 UNINITIALIZED_PROFILE
,
34 /* Profile is based on static branch prediction heuristics and may
35 or may not match reality. It is local to function and cannot be compared
36 inter-procedurally. Never used by probabilities (they are always local).
40 /* Profile was read by feedback and was 0, we used local heuristics to guess
41 better. This is the case of functions not run in profile feedback.
42 Never used by probabilities. */
45 /* Same as GUESSED_GLOBAL0 but global count is adjusted 0. */
46 GUESSED_GLOBAL0_ADJUSTED
,
48 /* Profile is based on static branch prediction heuristics. It may or may
49 not reflect the reality but it can be compared interprocedurally
50 (for example, we inlined function w/o profile feedback into function
51 with feedback and propagated from that).
52 Never used by probabilities. */
55 /* Profile was determined by autofdo. */
58 /* Profile was originally based on feedback but it was adjusted
59 by code duplicating optimization. It may not precisely reflect the
60 particular code path. */
63 /* Profile was read from profile feedback or determined by accurate static
68 extern const char *profile_quality_as_string (enum profile_quality
);
69 extern bool parse_profile_quality (const char *value
,
70 profile_quality
*quality
);
72 /* The base value for branch probability notes and edge probabilities. */
73 #define REG_BR_PROB_BASE 10000
75 #define RDIV(X,Y) (((X) + (Y) / 2) / (Y))
77 bool slow_safe_scale_64bit (uint64_t a
, uint64_t b
, uint64_t c
, uint64_t *res
);
79 /* Compute RES=(a*b + c/2)/c capping and return false if overflow happened. */
82 safe_scale_64bit (uint64_t a
, uint64_t b
, uint64_t c
, uint64_t *res
)
84 #if (GCC_VERSION >= 5000)
86 if (!__builtin_mul_overflow (a
, b
, &tmp
)
87 && !__builtin_add_overflow (tmp
, c
/2, &tmp
))
98 if (a
< ((uint64_t)1 << 31)
99 && b
< ((uint64_t)1 << 31)
100 && c
< ((uint64_t)1 << 31))
102 *res
= (a
* b
+ (c
/ 2)) / c
;
106 return slow_safe_scale_64bit (a
, b
, c
, res
);
109 /* Data type to hold probabilities. It implements fixed point arithmetics
110 with capping so probability is always in range [0,1] and scaling requiring
111 values greater than 1 needs to be represented otherwise.
113 In addition to actual value the quality of profile is tracked and propagated
114 through all operations. Special value UNINITIALIZED_PROFILE is used for probabilities
115 that has not been determined yet (for example because of
116 -fno-guess-branch-probability)
118 Typically probabilities are derived from profile feedback (via
119 probability_in_gcov_type), autoFDO or guessed statically and then propagated
120 thorough the compilation.
122 Named probabilities are available:
123 - never (0 probability)
125 - very_unlikely (1/2000 probability)
126 - unlikely (1/5 probability)
127 - even (1/2 probability)
128 - likely (4/5 probability)
129 - very_likely (1999/2000 probability)
133 Named probabilities except for never/always are assumed to be statically
134 guessed and thus not necessarily accurate. The difference between never
135 and guessed_never is that the first one should be used only in case that
136 well behaving program will very likely not execute the "never" path.
137 For example if the path is going to abort () call or it exception handling.
139 Always and guessed_always probabilities are symmetric.
141 For legacy code we support conversion to/from REG_BR_PROB_BASE based fixpoint
142 integer arithmetics. Once the code is converted to branch probabilities,
143 these conversions will probably go away because they are lossy.
146 class GTY((user
)) profile_probability
148 static const int n_bits
= 29;
149 /* We can technically use ((uint32_t) 1 << (n_bits - 1)) - 2 but that
150 will lead to harder multiplication sequences. */
151 static const uint32_t max_probability
= (uint32_t) 1 << (n_bits
- 2);
152 static const uint32_t uninitialized_probability
153 = ((uint32_t) 1 << (n_bits
- 1)) - 1;
156 enum profile_quality m_quality
: 3;
158 friend struct profile_count
;
160 profile_probability (): m_val (uninitialized_probability
),
164 profile_probability (uint32_t val
, profile_quality quality
):
165 m_val (val
), m_quality (quality
)
168 /* Named probabilities. */
169 static profile_probability
never ()
171 profile_probability ret
;
173 ret
.m_quality
= PRECISE
;
177 static profile_probability
guessed_never ()
179 profile_probability ret
;
181 ret
.m_quality
= GUESSED
;
185 static profile_probability
very_unlikely ()
187 /* Be consistent with PROB_VERY_UNLIKELY in predict.h. */
188 profile_probability r
= guessed_always () / 2000;
193 static profile_probability
unlikely ()
195 /* Be consistent with PROB_VERY_LIKELY in predict.h. */
196 profile_probability r
= guessed_always () / 5;
201 static profile_probability
even ()
203 return guessed_always () / 2;
206 static profile_probability
very_likely ()
208 return always () - very_unlikely ();
211 static profile_probability
likely ()
213 return always () - unlikely ();
216 static profile_probability
guessed_always ()
218 profile_probability ret
;
219 ret
.m_val
= max_probability
;
220 ret
.m_quality
= GUESSED
;
224 static profile_probability
always ()
226 profile_probability ret
;
227 ret
.m_val
= max_probability
;
228 ret
.m_quality
= PRECISE
;
232 /* Probabilities which has not been initialized. Either because
233 initialization did not happen yet or because profile is unknown. */
234 static profile_probability
uninitialized ()
236 profile_probability c
;
237 c
.m_val
= uninitialized_probability
;
238 c
.m_quality
= GUESSED
;
242 /* Return true if value has been initialized. */
243 bool initialized_p () const
245 return m_val
!= uninitialized_probability
;
248 /* Return true if value can be trusted. */
249 bool reliable_p () const
251 return m_quality
>= ADJUSTED
;
254 /* Conversion from and to REG_BR_PROB_BASE integer fixpoint arithmetics.
255 this is mostly to support legacy code and should go away. */
256 static profile_probability
from_reg_br_prob_base (int v
)
258 profile_probability ret
;
259 gcc_checking_assert (v
>= 0 && v
<= REG_BR_PROB_BASE
);
260 ret
.m_val
= RDIV (v
* (uint64_t) max_probability
, REG_BR_PROB_BASE
);
261 ret
.m_quality
= GUESSED
;
265 /* Return THIS with quality set to ADJUSTED. */
266 profile_probability
adjusted () const
268 profile_probability ret
= *this;
269 if (!initialized_p ())
271 ret
.m_quality
= ADJUSTED
;
275 int to_reg_br_prob_base () const
277 gcc_checking_assert (initialized_p ());
278 return RDIV (m_val
* (uint64_t) REG_BR_PROB_BASE
, max_probability
);
281 /* Conversion to and from RTL representation of profile probabilities. */
282 static profile_probability
from_reg_br_prob_note (int v
)
284 profile_probability ret
;
285 ret
.m_val
= ((unsigned int)v
) / 8;
286 ret
.m_quality
= (enum profile_quality
)(v
& 7);
290 int to_reg_br_prob_note () const
292 gcc_checking_assert (initialized_p ());
293 int ret
= m_val
* 8 + m_quality
;
294 gcc_checking_assert (from_reg_br_prob_note (ret
) == *this);
298 /* Return VAL1/VAL2. */
299 static profile_probability probability_in_gcov_type
300 (gcov_type val1
, gcov_type val2
)
302 profile_probability ret
;
303 gcc_checking_assert (val1
>= 0 && val2
> 0);
305 ret
.m_val
= max_probability
;
309 safe_scale_64bit (val1
, max_probability
, val2
, &tmp
);
310 gcc_checking_assert (tmp
<= max_probability
);
313 ret
.m_quality
= PRECISE
;
317 /* Basic operations. */
318 bool operator== (const profile_probability
&other
) const
320 return m_val
== other
.m_val
&& m_quality
== other
.m_quality
;
323 profile_probability
operator+ (const profile_probability
&other
) const
325 if (other
== never ())
327 if (*this == never ())
329 if (!initialized_p () || !other
.initialized_p ())
330 return uninitialized ();
332 profile_probability ret
;
333 ret
.m_val
= MIN ((uint32_t)(m_val
+ other
.m_val
), max_probability
);
334 ret
.m_quality
= MIN (m_quality
, other
.m_quality
);
338 profile_probability
&operator+= (const profile_probability
&other
)
340 if (other
== never ())
342 if (*this == never ())
347 if (!initialized_p () || !other
.initialized_p ())
348 return *this = uninitialized ();
351 m_val
= MIN ((uint32_t)(m_val
+ other
.m_val
), max_probability
);
352 m_quality
= MIN (m_quality
, other
.m_quality
);
357 profile_probability
operator- (const profile_probability
&other
) const
359 if (*this == never ()
360 || other
== never ())
362 if (!initialized_p () || !other
.initialized_p ())
363 return uninitialized ();
364 profile_probability ret
;
365 ret
.m_val
= m_val
>= other
.m_val
? m_val
- other
.m_val
: 0;
366 ret
.m_quality
= MIN (m_quality
, other
.m_quality
);
370 profile_probability
&operator-= (const profile_probability
&other
)
372 if (*this == never ()
373 || other
== never ())
375 if (!initialized_p () || !other
.initialized_p ())
376 return *this = uninitialized ();
379 m_val
= m_val
>= other
.m_val
? m_val
- other
.m_val
: 0;
380 m_quality
= MIN (m_quality
, other
.m_quality
);
385 profile_probability
operator* (const profile_probability
&other
) const
387 if (*this == never ()
388 || other
== never ())
390 if (!initialized_p () || !other
.initialized_p ())
391 return uninitialized ();
392 profile_probability ret
;
393 ret
.m_val
= RDIV ((uint64_t)m_val
* other
.m_val
, max_probability
);
394 ret
.m_quality
= MIN (MIN (m_quality
, other
.m_quality
), ADJUSTED
);
398 profile_probability
&operator*= (const profile_probability
&other
)
400 if (*this == never ()
401 || other
== never ())
402 return *this = never ();
403 if (!initialized_p () || !other
.initialized_p ())
404 return *this = uninitialized ();
407 m_val
= RDIV ((uint64_t)m_val
* other
.m_val
, max_probability
);
408 m_quality
= MIN (MIN (m_quality
, other
.m_quality
), ADJUSTED
);
413 profile_probability
operator/ (const profile_probability
&other
) const
415 if (*this == never ())
417 if (!initialized_p () || !other
.initialized_p ())
418 return uninitialized ();
419 profile_probability ret
;
420 /* If we get probability above 1, mark it as unreliable and return 1. */
421 if (m_val
>= other
.m_val
)
423 ret
.m_val
= max_probability
;
424 ret
.m_quality
= MIN (MIN (m_quality
, other
.m_quality
),
432 gcc_checking_assert (other
.m_val
);
433 ret
.m_val
= MIN (RDIV ((uint64_t)m_val
* max_probability
,
437 ret
.m_quality
= MIN (MIN (m_quality
, other
.m_quality
), ADJUSTED
);
441 profile_probability
&operator/= (const profile_probability
&other
)
443 if (*this == never ())
444 return *this = never ();
445 if (!initialized_p () || !other
.initialized_p ())
446 return *this = uninitialized ();
449 /* If we get probability above 1, mark it as unreliable
451 if (m_val
> other
.m_val
)
453 m_val
= max_probability
;
454 m_quality
= MIN (MIN (m_quality
, other
.m_quality
),
462 gcc_checking_assert (other
.m_val
);
463 m_val
= MIN (RDIV ((uint64_t)m_val
* max_probability
,
467 m_quality
= MIN (MIN (m_quality
, other
.m_quality
), ADJUSTED
);
472 /* Split *THIS (ORIG) probability into 2 probabilities, such that
473 the returned one (FIRST) is *THIS * CPROB and *THIS is
474 adjusted (SECOND) so that FIRST + FIRST.invert () * SECOND
475 == ORIG. This is useful e.g. when splitting a conditional
478 goto lab; // ORIG probability
481 goto lab; // FIRST = ORIG * CPROB probability
483 goto lab; // SECOND probability
484 such that the overall probability of jumping to lab remains
485 the same. CPROB gives the relative probability between the
487 profile_probability
split (const profile_probability
&cprob
)
489 profile_probability ret
= *this * cprob
;
490 /* The following is equivalent to:
491 *this = cprob.invert () * *this / ret.invert ();
492 Avoid scaling when overall outcome is supposed to be always.
493 Without knowing that one is inverse of other, the result would be
495 if (!(*this == always ()))
496 *this = (*this - ret
) / ret
.invert ();
500 gcov_type
apply (gcov_type val
) const
502 if (*this == uninitialized ())
504 return RDIV (val
* m_val
, max_probability
);
507 /* Return 1-*THIS. */
508 profile_probability
invert () const
510 return always() - *this;
513 /* Return THIS with quality dropped to GUESSED. */
514 profile_probability
guessed () const
516 profile_probability ret
= *this;
517 ret
.m_quality
= GUESSED
;
521 /* Return THIS with quality dropped to AFDO. */
522 profile_probability
afdo () const
524 profile_probability ret
= *this;
525 ret
.m_quality
= AFDO
;
529 /* Return *THIS * NUM / DEN. */
530 profile_probability
apply_scale (int64_t num
, int64_t den
) const
532 if (*this == never ())
534 if (!initialized_p ())
535 return uninitialized ();
536 profile_probability ret
;
538 safe_scale_64bit (m_val
, num
, den
, &tmp
);
539 ret
.m_val
= MIN (tmp
, max_probability
);
540 ret
.m_quality
= MIN (m_quality
, ADJUSTED
);
544 /* Return true when the probability of edge is reliable.
546 The profile guessing code is good at predicting branch outcome (i.e.
547 taken/not taken), that is predicted right slightly over 75% of time.
548 It is however notoriously poor on predicting the probability itself.
549 In general the profile appear a lot flatter (with probabilities closer
550 to 50%) than the reality so it is bad idea to use it to drive optimization
551 such as those disabling dynamic branch prediction for well predictable
554 There are two exceptions - edges leading to noreturn edges and edges
555 predicted by number of iterations heuristics are predicted well. This macro
556 should be able to distinguish those, but at the moment it simply check for
557 noreturn heuristic that is only one giving probability over 99% or bellow
558 1%. In future we might want to propagate reliability information across the
559 CFG if we find this information useful on multiple places. */
560 bool probably_reliable_p () const
562 if (m_quality
>= ADJUSTED
)
564 if (!initialized_p ())
566 return m_val
< max_probability
/ 100
567 || m_val
> max_probability
- max_probability
/ 100;
570 /* Return false if profile_probability is bogus. */
573 gcc_checking_assert (m_quality
!= UNINITIALIZED_PROFILE
);
574 if (m_val
== uninitialized_probability
)
575 return m_quality
== GUESSED
;
576 else if (m_quality
< GUESSED
)
578 return m_val
<= max_probability
;
581 /* Comparisons are three-state and conservative. False is returned if
582 the inequality cannot be decided. */
583 bool operator< (const profile_probability
&other
) const
585 return initialized_p () && other
.initialized_p () && m_val
< other
.m_val
;
588 bool operator> (const profile_probability
&other
) const
590 return initialized_p () && other
.initialized_p () && m_val
> other
.m_val
;
593 bool operator<= (const profile_probability
&other
) const
595 return initialized_p () && other
.initialized_p () && m_val
<= other
.m_val
;
598 bool operator>= (const profile_probability
&other
) const
600 return initialized_p () && other
.initialized_p () && m_val
>= other
.m_val
;
603 profile_probability
operator* (int64_t num
) const
605 return apply_scale (num
, 1);
608 profile_probability
operator*= (int64_t num
)
610 *this = apply_scale (num
, 1);
614 profile_probability
operator/ (int64_t den
) const
616 return apply_scale (1, den
);
619 profile_probability
operator/= (int64_t den
)
621 *this = apply_scale (1, den
);
625 /* Get the value of the count. */
626 uint32_t value () const { return m_val
; }
628 /* Get the quality of the count. */
629 enum profile_quality
quality () const { return m_quality
; }
631 /* Output THIS to F. */
632 void dump (FILE *f
) const;
634 /* Output THIS to BUFFER. */
635 void dump (char *buffer
) const;
637 /* Print THIS to stderr. */
640 /* Return true if THIS is known to differ significantly from OTHER. */
641 bool differs_from_p (profile_probability other
) const;
643 /* Return if difference is greater than 50%. */
644 bool differs_lot_from_p (profile_probability other
) const;
646 /* COUNT1 times event happens with *THIS probability, COUNT2 times OTHER
647 happens with COUNT2 probability. Return probability that either *THIS or
649 profile_probability
combine_with_count (profile_count count1
,
650 profile_probability other
,
651 profile_count count2
) const;
653 /* Return probability as sreal. */
654 sreal
to_sreal () const;
655 /* LTO streaming support. */
656 static profile_probability
stream_in (class lto_input_block
*);
657 void stream_out (struct output_block
*);
658 void stream_out (struct lto_output_stream
*);
661 /* Main data type to hold profile counters in GCC. Profile counts originate
662 either from profile feedback, static profile estimation or both. We do not
663 perform whole program profile propagation and thus profile estimation
664 counters are often local to function, while counters from profile feedback
665 (or special cases of profile estimation) can be used inter-procedurally.
667 There are 3 basic types
668 1) local counters which are result of intra-procedural static profile
670 2) ipa counters which are result of profile feedback or special case
671 of static profile estimation (such as in function main).
672 3) counters which counts as 0 inter-procedurally (because given function
673 was never run in train feedback) but they hold local static profile
676 Counters of type 1 and 3 cannot be mixed with counters of different type
677 within operation (because whole function should use one type of counter)
678 with exception that global zero mix in most operations where outcome is
681 To take local counter and use it inter-procedurally use ipa member function
682 which strips information irrelevant at the inter-procedural level.
684 Counters are 61bit integers representing number of executions during the
685 train run or normalized frequency within the function.
687 As the profile is maintained during the compilation, many adjustments are
688 made. Not all transformations can be made precisely, most importantly
689 when code is being duplicated. It also may happen that part of CFG has
690 profile counts known while other do not - for example when LTO optimizing
691 partly profiled program or when profile was lost due to COMDAT merging.
693 For this reason profile_count tracks more information than
694 just unsigned integer and it is also ready for profile mismatches.
695 The API of this data type represent operations that are natural
696 on profile counts - sum, difference and operation with scales and
697 probabilities. All operations are safe by never getting negative counts
698 and they do end up in uninitialized scale if any of the parameters is
701 All comparisons that are three state and handling of probabilities. Thus
702 a < b is not equal to !(a >= b).
704 The following pre-defined counts are available:
706 profile_count::zero () for code that is known to execute zero times at
707 runtime (this can be detected statically i.e. for paths leading to
709 profile_count::one () for code that is known to execute once (such as
711 profile_count::uninitialized () for unknown execution count.
715 struct GTY(()) profile_count
718 /* Use 62bit to hold basic block counters. Should be at least
719 64bit. Although a counter cannot be negative, we use a signed
720 type to hold various extra stages. */
722 static const int n_bits
= 61;
723 static const uint64_t max_count
= ((uint64_t) 1 << n_bits
) - 2;
725 static const uint64_t uninitialized_count
= ((uint64_t) 1 << n_bits
) - 1;
727 #if defined (__arm__) && (__GNUC__ >= 6 && __GNUC__ <= 8)
728 /* Work-around for PR88469. A bug in the gcc-6/7/8 PCS layout code
729 incorrectly detects the alignment of a structure where the only
730 64-bit aligned object is a bit-field. We force the alignment of
731 the entire field to mitigate this. */
732 #define UINT64_BIT_FIELD_ALIGN __attribute__ ((aligned(8)))
734 #define UINT64_BIT_FIELD_ALIGN
736 uint64_t UINT64_BIT_FIELD_ALIGN m_val
: n_bits
;
737 #undef UINT64_BIT_FIELD_ALIGN
738 enum profile_quality m_quality
: 3;
741 /* Return true if both values can meaningfully appear in single function
742 body. We have either all counters in function local or global, otherwise
743 operations between them are not really defined well. */
744 bool compatible_p (const profile_count other
) const
746 if (!initialized_p () || !other
.initialized_p ())
751 /* Do not allow nonzero global profile together with local guesses
752 that are globally0. */
753 if (ipa ().nonzero_p ()
754 && !(other
.ipa () == other
))
756 if (other
.ipa ().nonzero_p ()
757 && !(ipa () == *this))
760 return ipa_p () == other
.ipa_p ();
763 /* Used for counters which are expected to be never executed. */
764 static profile_count
zero ()
766 return from_gcov_type (0);
769 static profile_count
adjusted_zero ()
773 c
.m_quality
= ADJUSTED
;
777 static profile_count
guessed_zero ()
781 c
.m_quality
= GUESSED
;
785 static profile_count
one ()
787 return from_gcov_type (1);
790 /* Value of counters which has not been initialized. Either because
791 initialization did not happen yet or because profile is unknown. */
792 static profile_count
uninitialized ()
795 c
.m_val
= uninitialized_count
;
796 c
.m_quality
= GUESSED_LOCAL
;
800 /* Conversion to gcov_type is lossy. */
801 gcov_type
to_gcov_type () const
803 gcc_checking_assert (initialized_p ());
807 /* Return true if value has been initialized. */
808 bool initialized_p () const
810 return m_val
!= uninitialized_count
;
813 /* Return true if value can be trusted. */
814 bool reliable_p () const
816 return m_quality
>= ADJUSTED
;
819 /* Return true if value can be operated inter-procedurally. */
822 return !initialized_p () || m_quality
>= GUESSED_GLOBAL0
;
825 /* Return true if quality of profile is precise. */
826 bool precise_p () const
828 return m_quality
== PRECISE
;
831 /* Get the value of the count. */
832 uint64_t value () const { return m_val
; }
834 /* Get the quality of the count. */
835 enum profile_quality
quality () const { return m_quality
; }
837 /* When merging basic blocks, the two different profile counts are unified.
838 Return true if this can be done without losing info about profile.
839 The only case we care about here is when first BB contains something
840 that makes it terminate in a way not visible in CFG. */
841 bool ok_for_merging (profile_count other
) const
843 if (m_quality
< ADJUSTED
844 || other
.m_quality
< ADJUSTED
)
846 return !(other
< *this);
849 /* When merging two BBs with different counts, pick common count that looks
850 most representative. */
851 profile_count
merge (profile_count other
) const
853 if (*this == other
|| !other
.initialized_p ()
854 || m_quality
> other
.m_quality
)
856 if (other
.m_quality
> m_quality
862 /* Basic operations. */
863 bool operator== (const profile_count
&other
) const
865 return m_val
== other
.m_val
&& m_quality
== other
.m_quality
;
868 profile_count
operator+ (const profile_count
&other
) const
870 if (other
== zero ())
872 if (*this == zero ())
874 if (!initialized_p () || !other
.initialized_p ())
875 return uninitialized ();
878 gcc_checking_assert (compatible_p (other
));
879 ret
.m_val
= m_val
+ other
.m_val
;
880 ret
.m_quality
= MIN (m_quality
, other
.m_quality
);
884 profile_count
&operator+= (const profile_count
&other
)
886 if (other
== zero ())
888 if (*this == zero ())
893 if (!initialized_p () || !other
.initialized_p ())
894 return *this = uninitialized ();
897 gcc_checking_assert (compatible_p (other
));
898 m_val
+= other
.m_val
;
899 m_quality
= MIN (m_quality
, other
.m_quality
);
904 profile_count
operator- (const profile_count
&other
) const
906 if (*this == zero () || other
== zero ())
908 if (!initialized_p () || !other
.initialized_p ())
909 return uninitialized ();
910 gcc_checking_assert (compatible_p (other
));
912 ret
.m_val
= m_val
>= other
.m_val
? m_val
- other
.m_val
: 0;
913 ret
.m_quality
= MIN (m_quality
, other
.m_quality
);
917 profile_count
&operator-= (const profile_count
&other
)
919 if (*this == zero () || other
== zero ())
921 if (!initialized_p () || !other
.initialized_p ())
922 return *this = uninitialized ();
925 gcc_checking_assert (compatible_p (other
));
926 m_val
= m_val
>= other
.m_val
? m_val
- other
.m_val
: 0;
927 m_quality
= MIN (m_quality
, other
.m_quality
);
932 /* Return false if profile_count is bogus. */
935 gcc_checking_assert (m_quality
!= UNINITIALIZED_PROFILE
);
936 return m_val
!= uninitialized_count
|| m_quality
== GUESSED_LOCAL
;
939 /* Comparisons are three-state and conservative. False is returned if
940 the inequality cannot be decided. */
941 bool operator< (const profile_count
&other
) const
943 if (!initialized_p () || !other
.initialized_p ())
945 if (*this == zero ())
946 return !(other
== zero ());
947 if (other
== zero ())
949 gcc_checking_assert (compatible_p (other
));
950 return m_val
< other
.m_val
;
953 bool operator> (const profile_count
&other
) const
955 if (!initialized_p () || !other
.initialized_p ())
957 if (*this == zero ())
959 if (other
== zero ())
960 return !(*this == zero ());
961 gcc_checking_assert (compatible_p (other
));
962 return initialized_p () && other
.initialized_p () && m_val
> other
.m_val
;
965 bool operator< (const gcov_type other
) const
967 gcc_checking_assert (ipa_p ());
968 gcc_checking_assert (other
>= 0);
969 return ipa ().initialized_p () && ipa ().m_val
< (uint64_t) other
;
972 bool operator> (const gcov_type other
) const
974 gcc_checking_assert (ipa_p ());
975 gcc_checking_assert (other
>= 0);
976 return ipa ().initialized_p () && ipa ().m_val
> (uint64_t) other
;
979 bool operator<= (const profile_count
&other
) const
981 if (!initialized_p () || !other
.initialized_p ())
983 if (*this == zero ())
985 if (other
== zero ())
986 return (*this == zero ());
987 gcc_checking_assert (compatible_p (other
));
988 return m_val
<= other
.m_val
;
991 bool operator>= (const profile_count
&other
) const
993 if (!initialized_p () || !other
.initialized_p ())
995 if (other
== zero ())
997 if (*this == zero ())
998 return (other
== zero ());
999 gcc_checking_assert (compatible_p (other
));
1000 return m_val
>= other
.m_val
;
1003 bool operator<= (const gcov_type other
) const
1005 gcc_checking_assert (ipa_p ());
1006 gcc_checking_assert (other
>= 0);
1007 return ipa ().initialized_p () && ipa ().m_val
<= (uint64_t) other
;
1010 bool operator>= (const gcov_type other
) const
1012 gcc_checking_assert (ipa_p ());
1013 gcc_checking_assert (other
>= 0);
1014 return ipa ().initialized_p () && ipa ().m_val
>= (uint64_t) other
;
1017 profile_count
operator* (int64_t num
) const
1019 return apply_scale (num
, 1);
1022 profile_count
operator*= (int64_t num
)
1024 *this = apply_scale (num
, 1);
1028 profile_count
operator/ (int64_t den
) const
1030 return apply_scale (1, den
);
1033 profile_count
operator/= (int64_t den
)
1035 *this = apply_scale (1, den
);
1039 /* Return true when value is not zero and can be used for scaling.
1040 This is different from *this > 0 because that requires counter to
1042 bool nonzero_p () const
1044 return initialized_p () && m_val
!= 0;
1047 /* Make counter forcibly nonzero. */
1048 profile_count
force_nonzero () const
1050 if (!initialized_p ())
1052 profile_count ret
= *this;
1056 ret
.m_quality
= MIN (m_quality
, ADJUSTED
);
1061 profile_count
max (profile_count other
) const
1063 profile_count val
= *this;
1065 /* Always prefer nonzero IPA counts over local counts. */
1066 if (ipa ().nonzero_p () || other
.ipa ().nonzero_p ())
1069 other
= other
.ipa ();
1071 if (!initialized_p ())
1073 if (!other
.initialized_p ())
1075 if (*this == zero ())
1077 if (other
== zero ())
1079 gcc_checking_assert (compatible_p (other
));
1080 if (val
.m_val
< other
.m_val
|| (m_val
== other
.m_val
1081 && val
.m_quality
< other
.m_quality
))
1086 /* PROB is a probability in scale 0...REG_BR_PROB_BASE. Scale counter
1088 profile_count
apply_probability (int prob
) const
1090 gcc_checking_assert (prob
>= 0 && prob
<= REG_BR_PROB_BASE
);
1093 if (!initialized_p ())
1094 return uninitialized ();
1096 ret
.m_val
= RDIV (m_val
* prob
, REG_BR_PROB_BASE
);
1097 ret
.m_quality
= MIN (m_quality
, ADJUSTED
);
1101 /* Scale counter according to PROB. */
1102 profile_count
apply_probability (profile_probability prob
) const
1104 if (*this == zero ())
1106 if (prob
== profile_probability::never ())
1108 if (!initialized_p ())
1109 return uninitialized ();
1112 safe_scale_64bit (m_val
, prob
.m_val
, profile_probability::max_probability
,
1115 ret
.m_quality
= MIN (m_quality
, prob
.m_quality
);
1119 /* Return *THIS * NUM / DEN. */
1120 profile_count
apply_scale (int64_t num
, int64_t den
) const
1124 if (!initialized_p ())
1125 return uninitialized ();
1129 gcc_checking_assert (num
>= 0 && den
> 0);
1130 safe_scale_64bit (m_val
, num
, den
, &tmp
);
1131 ret
.m_val
= MIN (tmp
, max_count
);
1132 ret
.m_quality
= MIN (m_quality
, ADJUSTED
);
1136 profile_count
apply_scale (profile_count num
, profile_count den
) const
1138 if (*this == zero ())
1142 if (!initialized_p () || !num
.initialized_p () || !den
.initialized_p ())
1143 return uninitialized ();
1146 gcc_checking_assert (den
.m_val
);
1150 safe_scale_64bit (m_val
, num
.m_val
, den
.m_val
, &val
);
1151 ret
.m_val
= MIN (val
, max_count
);
1152 ret
.m_quality
= MIN (MIN (MIN (m_quality
, ADJUSTED
),
1153 num
.m_quality
), den
.m_quality
);
1154 /* Be sure that ret is not local if num is global.
1155 Also ensure that ret is not global0 when num is global. */
1157 ret
.m_quality
= MAX (ret
.m_quality
,
1158 num
== num
.ipa () ? GUESSED
: num
.m_quality
);
1162 /* Return THIS with quality dropped to GUESSED_LOCAL. */
1163 profile_count
guessed_local () const
1165 profile_count ret
= *this;
1166 if (!initialized_p ())
1168 ret
.m_quality
= GUESSED_LOCAL
;
1172 /* We know that profile is globally 0 but keep local profile if present. */
1173 profile_count
global0 () const
1175 profile_count ret
= *this;
1176 if (!initialized_p ())
1178 ret
.m_quality
= GUESSED_GLOBAL0
;
1182 /* We know that profile is globally adjusted 0 but keep local profile
1184 profile_count
global0adjusted () const
1186 profile_count ret
= *this;
1187 if (!initialized_p ())
1189 ret
.m_quality
= GUESSED_GLOBAL0_ADJUSTED
;
1193 /* Return THIS with quality dropped to GUESSED. */
1194 profile_count
guessed () const
1196 profile_count ret
= *this;
1197 ret
.m_quality
= MIN (ret
.m_quality
, GUESSED
);
1201 /* Return variant of profile count which is always safe to compare
1202 across functions. */
1203 profile_count
ipa () const
1205 if (m_quality
> GUESSED_GLOBAL0_ADJUSTED
)
1207 if (m_quality
== GUESSED_GLOBAL0
)
1209 if (m_quality
== GUESSED_GLOBAL0_ADJUSTED
)
1210 return adjusted_zero ();
1211 return uninitialized ();
1214 /* Return THIS with quality dropped to AFDO. */
1215 profile_count
afdo () const
1217 profile_count ret
= *this;
1218 ret
.m_quality
= AFDO
;
1222 /* Return probability of event with counter THIS within event with counter
1224 profile_probability
probability_in (const profile_count overall
) const
1226 if (*this == zero ()
1227 && !(overall
== zero ()))
1228 return profile_probability::never ();
1229 if (!initialized_p () || !overall
.initialized_p ()
1231 return profile_probability::uninitialized ();
1232 if (*this == overall
&& m_quality
== PRECISE
)
1233 return profile_probability::always ();
1234 profile_probability ret
;
1235 gcc_checking_assert (compatible_p (overall
));
1237 if (overall
.m_val
< m_val
)
1239 ret
.m_val
= profile_probability::max_probability
;
1240 ret
.m_quality
= GUESSED
;
1244 ret
.m_val
= RDIV (m_val
* profile_probability::max_probability
,
1246 ret
.m_quality
= MIN (MAX (MIN (m_quality
, overall
.m_quality
),
1247 GUESSED
), ADJUSTED
);
1251 int to_frequency (struct function
*fun
) const;
1252 int to_cgraph_frequency (profile_count entry_bb_count
) const;
1253 sreal
to_sreal_scale (profile_count in
, bool *known
= NULL
) const;
1255 /* Output THIS to F. */
1256 void dump (FILE *f
) const;
1258 /* Output THIS to BUFFER. */
1259 void dump (char *buffer
) const;
1261 /* Print THIS to stderr. */
1262 void debug () const;
1264 /* Return true if THIS is known to differ significantly from OTHER. */
1265 bool differs_from_p (profile_count other
) const;
1267 /* We want to scale profile across function boundary from NUM to DEN.
1268 Take care of the side case when NUM and DEN are zeros of incompatible
1270 static void adjust_for_ipa_scaling (profile_count
*num
, profile_count
*den
);
1272 /* THIS is a count of bb which is known to be executed IPA times.
1273 Combine this information into bb counter. This means returning IPA
1274 if it is nonzero, not changing anything if IPA is uninitialized
1275 and if IPA is zero, turning THIS into corresponding local profile with
1277 profile_count
combine_with_ipa_count (profile_count ipa
);
1279 /* Same as combine_with_ipa_count but inside function with count IPA2. */
1280 profile_count combine_with_ipa_count_within
1281 (profile_count ipa
, profile_count ipa2
);
1283 /* The profiling runtime uses gcov_type, which is usually 64bit integer.
1284 Conversions back and forth are used to read the coverage and get it
1285 into internal representation. */
1286 static profile_count
from_gcov_type (gcov_type v
,
1287 profile_quality quality
= PRECISE
);
1289 /* LTO streaming support. */
1290 static profile_count
stream_in (class lto_input_block
*);
1291 void stream_out (struct output_block
*);
1292 void stream_out (struct lto_output_stream
*);