From e5392ef25735dbf5246ec00c962f8523ebd89aef Mon Sep 17 00:00:00 2001 From: rsandifo Date: Wed, 30 Aug 2017 11:08:00 +0000 Subject: [PATCH] Split out parts of scompare_loc_descriptor and emit_store_flag This patch splits some cases out of scompare_loc_descriptor and emit_store_flag, which helps with the upcoming machmode series. 2017-08-30 Richard Sandiford gcc/ * dwarf2out.c (scompare_loc_descriptor_wide) (scompare_loc_descriptor_narrow): New functions, split out from... (scompare_loc_descriptor): ...here. * expmed.c (emit_store_flag_int): New function, split out from... (emit_store_flag): ...here. git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@251451 138bc75d-0d04-0410-961f-82ee72b054a4 --- gcc/ChangeLog | 8 ++ gcc/dwarf2out.c | 116 ++++++++++++--------- gcc/expmed.c | 319 ++++++++++++++++++++++++++++++-------------------------- 3 files changed, 246 insertions(+), 197 deletions(-) diff --git a/gcc/ChangeLog b/gcc/ChangeLog index 0aa097f2b57..64aa109c4ac 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,3 +1,11 @@ +2017-08-30 Richard Sandiford + + * dwarf2out.c (scompare_loc_descriptor_wide) + (scompare_loc_descriptor_narrow): New functions, split out from... + (scompare_loc_descriptor): ...here. + * expmed.c (emit_store_flag_int): New function, split out from... + (emit_store_flag): ...here. + 2017-08-30 Richard Biener * dwarf2out.c (dwarf2out_finish): Remove setting AT_pubnames. diff --git a/gcc/dwarf2out.c b/gcc/dwarf2out.c index 00aab04f498..e17b58ab7b0 100644 --- a/gcc/dwarf2out.c +++ b/gcc/dwarf2out.c @@ -13929,60 +13929,43 @@ compare_loc_descriptor (enum dwarf_location_atom op, dw_loc_descr_ref op0, return ret; } -/* Return location descriptor for signed comparison OP RTL. */ +/* Subroutine of scompare_loc_descriptor for the case in which we're + comparing two scalar integer operands OP0 and OP1 that have mode OP_MODE, + and in which OP_MODE is bigger than DWARF2_ADDR_SIZE. */ static dw_loc_descr_ref -scompare_loc_descriptor (enum dwarf_location_atom op, rtx rtl, - machine_mode mem_mode) +scompare_loc_descriptor_wide (enum dwarf_location_atom op, + machine_mode op_mode, + dw_loc_descr_ref op0, dw_loc_descr_ref op1) { - machine_mode op_mode = GET_MODE (XEXP (rtl, 0)); - dw_loc_descr_ref op0, op1; - int shift; - - if (op_mode == VOIDmode) - op_mode = GET_MODE (XEXP (rtl, 1)); - if (op_mode == VOIDmode) - return NULL; - - if (dwarf_strict - && dwarf_version < 5 - && (!SCALAR_INT_MODE_P (op_mode) - || GET_MODE_SIZE (op_mode) > DWARF2_ADDR_SIZE)) - return NULL; - - op0 = mem_loc_descriptor (XEXP (rtl, 0), op_mode, mem_mode, - VAR_INIT_STATUS_INITIALIZED); - op1 = mem_loc_descriptor (XEXP (rtl, 1), op_mode, mem_mode, - VAR_INIT_STATUS_INITIALIZED); + dw_die_ref type_die = base_type_for_mode (op_mode, 0); + dw_loc_descr_ref cvt; - if (op0 == NULL || op1 == NULL) + if (type_die == NULL) return NULL; + cvt = new_loc_descr (dwarf_OP (DW_OP_convert), 0, 0); + cvt->dw_loc_oprnd1.val_class = dw_val_class_die_ref; + cvt->dw_loc_oprnd1.v.val_die_ref.die = type_die; + cvt->dw_loc_oprnd1.v.val_die_ref.external = 0; + add_loc_descr (&op0, cvt); + cvt = new_loc_descr (dwarf_OP (DW_OP_convert), 0, 0); + cvt->dw_loc_oprnd1.val_class = dw_val_class_die_ref; + cvt->dw_loc_oprnd1.v.val_die_ref.die = type_die; + cvt->dw_loc_oprnd1.v.val_die_ref.external = 0; + add_loc_descr (&op1, cvt); + return compare_loc_descriptor (op, op0, op1); +} - if (!SCALAR_INT_MODE_P (op_mode) - || GET_MODE_SIZE (op_mode) == DWARF2_ADDR_SIZE) - return compare_loc_descriptor (op, op0, op1); - - if (GET_MODE_SIZE (op_mode) > DWARF2_ADDR_SIZE) - { - dw_die_ref type_die = base_type_for_mode (op_mode, 0); - dw_loc_descr_ref cvt; +/* Subroutine of scompare_loc_descriptor for the case in which we're + comparing two scalar integer operands OP0 and OP1 that have mode OP_MODE, + and in which OP_MODE is smaller than DWARF2_ADDR_SIZE. */ - if (type_die == NULL) - return NULL; - cvt = new_loc_descr (dwarf_OP (DW_OP_convert), 0, 0); - cvt->dw_loc_oprnd1.val_class = dw_val_class_die_ref; - cvt->dw_loc_oprnd1.v.val_die_ref.die = type_die; - cvt->dw_loc_oprnd1.v.val_die_ref.external = 0; - add_loc_descr (&op0, cvt); - cvt = new_loc_descr (dwarf_OP (DW_OP_convert), 0, 0); - cvt->dw_loc_oprnd1.val_class = dw_val_class_die_ref; - cvt->dw_loc_oprnd1.v.val_die_ref.die = type_die; - cvt->dw_loc_oprnd1.v.val_die_ref.external = 0; - add_loc_descr (&op1, cvt); - return compare_loc_descriptor (op, op0, op1); - } - - shift = (DWARF2_ADDR_SIZE - GET_MODE_SIZE (op_mode)) * BITS_PER_UNIT; +static dw_loc_descr_ref +scompare_loc_descriptor_narrow (enum dwarf_location_atom op, rtx rtl, + machine_mode op_mode, + dw_loc_descr_ref op0, dw_loc_descr_ref op1) +{ + int shift = (DWARF2_ADDR_SIZE - GET_MODE_SIZE (op_mode)) * BITS_PER_UNIT; /* For eq/ne, if the operands are known to be zero-extended, there is no need to do the fancy shifting up. */ if (op == DW_OP_eq || op == DW_OP_ne) @@ -14040,6 +14023,45 @@ scompare_loc_descriptor (enum dwarf_location_atom op, rtx rtl, return compare_loc_descriptor (op, op0, op1); } +/* Return location descriptor for signed comparison OP RTL. */ + +static dw_loc_descr_ref +scompare_loc_descriptor (enum dwarf_location_atom op, rtx rtl, + machine_mode mem_mode) +{ + machine_mode op_mode = GET_MODE (XEXP (rtl, 0)); + dw_loc_descr_ref op0, op1; + + if (op_mode == VOIDmode) + op_mode = GET_MODE (XEXP (rtl, 1)); + if (op_mode == VOIDmode) + return NULL; + + if (dwarf_strict + && dwarf_version < 5 + && (!SCALAR_INT_MODE_P (op_mode) + || GET_MODE_SIZE (op_mode) > DWARF2_ADDR_SIZE)) + return NULL; + + op0 = mem_loc_descriptor (XEXP (rtl, 0), op_mode, mem_mode, + VAR_INIT_STATUS_INITIALIZED); + op1 = mem_loc_descriptor (XEXP (rtl, 1), op_mode, mem_mode, + VAR_INIT_STATUS_INITIALIZED); + + if (op0 == NULL || op1 == NULL) + return NULL; + + if (SCALAR_INT_MODE_P (op_mode)) + { + if (GET_MODE_SIZE (op_mode) < DWARF2_ADDR_SIZE) + return scompare_loc_descriptor_narrow (op, rtl, op_mode, op0, op1); + + if (GET_MODE_SIZE (op_mode) > DWARF2_ADDR_SIZE) + return scompare_loc_descriptor_wide (op, op_mode, op0, op1); + } + return compare_loc_descriptor (op, op0, op1); +} + /* Return location descriptor for unsigned comparison OP RTL. */ static dw_loc_descr_ref diff --git a/gcc/expmed.c b/gcc/expmed.c index 37f2df819fb..7eeadc1d694 100644 --- a/gcc/expmed.c +++ b/gcc/expmed.c @@ -5583,155 +5583,19 @@ emit_store_flag_1 (rtx target, enum rtx_code code, rtx op0, rtx op1, return 0; } -/* Emit a store-flags instruction for comparison CODE on OP0 and OP1 - and storing in TARGET. Normally return TARGET. - Return 0 if that cannot be done. - - MODE is the mode to use for OP0 and OP1 should they be CONST_INTs. If - it is VOIDmode, they cannot both be CONST_INT. - - UNSIGNEDP is for the case where we have to widen the operands - to perform the operation. It says to use zero-extension. - - NORMALIZEP is 1 if we should convert the result to be either zero - or one. Normalize is -1 if we should convert the result to be - either zero or -1. If NORMALIZEP is zero, the result will be left - "raw" out of the scc insn. */ +/* Subroutine of emit_store_flag that handles cases in which the operands + are scalar integers. SUBTARGET is the target to use for temporary + operations and TRUEVAL is the value to store when the condition is + true. All other arguments are as for emit_store_flag. */ rtx -emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, - machine_mode mode, int unsignedp, int normalizep) +emit_store_flag_int (rtx target, rtx subtarget, enum rtx_code code, rtx op0, + rtx op1, machine_mode mode, int unsignedp, + int normalizep, rtx trueval) { machine_mode target_mode = target ? GET_MODE (target) : VOIDmode; - enum rtx_code rcode; - rtx subtarget; - rtx tem, trueval; - rtx_insn *last; - - /* If we compare constants, we shouldn't use a store-flag operation, - but a constant load. We can get there via the vanilla route that - usually generates a compare-branch sequence, but will in this case - fold the comparison to a constant, and thus elide the branch. */ - if (CONSTANT_P (op0) && CONSTANT_P (op1)) - return NULL_RTX; - - tem = emit_store_flag_1 (target, code, op0, op1, mode, unsignedp, normalizep, - target_mode); - if (tem) - return tem; - - /* If we reached here, we can't do this with a scc insn, however there - are some comparisons that can be done in other ways. Don't do any - of these cases if branches are very cheap. */ - if (BRANCH_COST (optimize_insn_for_speed_p (), false) == 0) - return 0; - - /* See what we need to return. We can only return a 1, -1, or the - sign bit. */ - - if (normalizep == 0) - { - if (STORE_FLAG_VALUE == 1 || STORE_FLAG_VALUE == -1) - normalizep = STORE_FLAG_VALUE; - - else if (val_signbit_p (mode, STORE_FLAG_VALUE)) - ; - else - return 0; - } - - last = get_last_insn (); - - /* If optimizing, use different pseudo registers for each insn, instead - of reusing the same pseudo. This leads to better CSE, but slows - down the compiler, since there are more pseudos */ - subtarget = (!optimize - && (target_mode == mode)) ? target : NULL_RTX; - trueval = GEN_INT (normalizep ? normalizep : STORE_FLAG_VALUE); - - /* For floating-point comparisons, try the reverse comparison or try - changing the "orderedness" of the comparison. */ - if (GET_MODE_CLASS (mode) == MODE_FLOAT) - { - enum rtx_code first_code; - bool and_them; - - rcode = reverse_condition_maybe_unordered (code); - if (can_compare_p (rcode, mode, ccp_store_flag) - && (code == ORDERED || code == UNORDERED - || (! HONOR_NANS (mode) && (code == LTGT || code == UNEQ)) - || (! HONOR_SNANS (mode) && (code == EQ || code == NE)))) - { - int want_add = ((STORE_FLAG_VALUE == 1 && normalizep == -1) - || (STORE_FLAG_VALUE == -1 && normalizep == 1)); - - /* For the reverse comparison, use either an addition or a XOR. */ - if (want_add - && rtx_cost (GEN_INT (normalizep), mode, PLUS, 1, - optimize_insn_for_speed_p ()) == 0) - { - tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, - STORE_FLAG_VALUE, target_mode); - if (tem) - return expand_binop (target_mode, add_optab, tem, - gen_int_mode (normalizep, target_mode), - target, 0, OPTAB_WIDEN); - } - else if (!want_add - && rtx_cost (trueval, mode, XOR, 1, - optimize_insn_for_speed_p ()) == 0) - { - tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, - normalizep, target_mode); - if (tem) - return expand_binop (target_mode, xor_optab, tem, trueval, - target, INTVAL (trueval) >= 0, OPTAB_WIDEN); - } - } - - delete_insns_since (last); - - /* Cannot split ORDERED and UNORDERED, only try the above trick. */ - if (code == ORDERED || code == UNORDERED) - return 0; - - and_them = split_comparison (code, mode, &first_code, &code); - - /* If there are no NaNs, the first comparison should always fall through. - Effectively change the comparison to the other one. */ - if (!HONOR_NANS (mode)) - { - gcc_assert (first_code == (and_them ? ORDERED : UNORDERED)); - return emit_store_flag_1 (target, code, op0, op1, mode, 0, normalizep, - target_mode); - } - - if (!HAVE_conditional_move) - return 0; - - /* Try using a setcc instruction for ORDERED/UNORDERED, followed by a - conditional move. */ - tem = emit_store_flag_1 (subtarget, first_code, op0, op1, mode, 0, - normalizep, target_mode); - if (tem == 0) - return 0; - - if (and_them) - tem = emit_conditional_move (target, code, op0, op1, mode, - tem, const0_rtx, GET_MODE (tem), 0); - else - tem = emit_conditional_move (target, code, op0, op1, mode, - trueval, tem, GET_MODE (tem), 0); - - if (tem == 0) - delete_insns_since (last); - return tem; - } - - /* The remaining tricks only apply to integer comparisons. */ - - if (GET_MODE_CLASS (mode) != MODE_INT) - return 0; + rtx_insn *last = get_last_insn (); + rtx tem; /* If this is an equality comparison of integers, we can try to exclusive-or (or subtract) the two operands and use a recursive call to try the @@ -5758,7 +5622,7 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, /* For integer comparisons, try the reverse comparison. However, for small X and if we'd have anyway to extend, implementing "X != 0" as "-(int)X >> 31" is still cheaper than inverting "(int)X == 0". */ - rcode = reverse_condition (code); + rtx_code rcode = reverse_condition (code); if (can_compare_p (rcode, mode, ccp_store_flag) && ! (optab_handler (cstore_optab, mode) == CODE_FOR_nothing && code == NE @@ -5776,7 +5640,7 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, STORE_FLAG_VALUE, target_mode); if (tem != 0) - tem = expand_binop (target_mode, add_optab, tem, + tem = expand_binop (target_mode, add_optab, tem, gen_int_mode (normalizep, target_mode), target, 0, OPTAB_WIDEN); } @@ -5787,7 +5651,7 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, normalizep, target_mode); if (tem != 0) - tem = expand_binop (target_mode, xor_optab, tem, trueval, target, + tem = expand_binop (target_mode, xor_optab, tem, trueval, target, INTVAL (trueval) >= 0, OPTAB_WIDEN); } @@ -5888,7 +5752,7 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, if (tem == 0 && (code == NE || BRANCH_COST (optimize_insn_for_speed_p (), - false) > 1)) + false) > 1)) { if (rtx_equal_p (subtarget, op0)) subtarget = 0; @@ -5910,7 +5774,7 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, if (tem) { if (!target) - ; + ; else if (GET_MODE (tem) != target_mode) { convert_move (target, tem, 0); @@ -5928,6 +5792,161 @@ emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, return tem; } +/* Emit a store-flags instruction for comparison CODE on OP0 and OP1 + and storing in TARGET. Normally return TARGET. + Return 0 if that cannot be done. + + MODE is the mode to use for OP0 and OP1 should they be CONST_INTs. If + it is VOIDmode, they cannot both be CONST_INT. + + UNSIGNEDP is for the case where we have to widen the operands + to perform the operation. It says to use zero-extension. + + NORMALIZEP is 1 if we should convert the result to be either zero + or one. Normalize is -1 if we should convert the result to be + either zero or -1. If NORMALIZEP is zero, the result will be left + "raw" out of the scc insn. */ + +rtx +emit_store_flag (rtx target, enum rtx_code code, rtx op0, rtx op1, + machine_mode mode, int unsignedp, int normalizep) +{ + machine_mode target_mode = target ? GET_MODE (target) : VOIDmode; + enum rtx_code rcode; + rtx subtarget; + rtx tem, trueval; + rtx_insn *last; + + /* If we compare constants, we shouldn't use a store-flag operation, + but a constant load. We can get there via the vanilla route that + usually generates a compare-branch sequence, but will in this case + fold the comparison to a constant, and thus elide the branch. */ + if (CONSTANT_P (op0) && CONSTANT_P (op1)) + return NULL_RTX; + + tem = emit_store_flag_1 (target, code, op0, op1, mode, unsignedp, normalizep, + target_mode); + if (tem) + return tem; + + /* If we reached here, we can't do this with a scc insn, however there + are some comparisons that can be done in other ways. Don't do any + of these cases if branches are very cheap. */ + if (BRANCH_COST (optimize_insn_for_speed_p (), false) == 0) + return 0; + + /* See what we need to return. We can only return a 1, -1, or the + sign bit. */ + + if (normalizep == 0) + { + if (STORE_FLAG_VALUE == 1 || STORE_FLAG_VALUE == -1) + normalizep = STORE_FLAG_VALUE; + + else if (val_signbit_p (mode, STORE_FLAG_VALUE)) + ; + else + return 0; + } + + last = get_last_insn (); + + /* If optimizing, use different pseudo registers for each insn, instead + of reusing the same pseudo. This leads to better CSE, but slows + down the compiler, since there are more pseudos. */ + subtarget = (!optimize + && (target_mode == mode)) ? target : NULL_RTX; + trueval = GEN_INT (normalizep ? normalizep : STORE_FLAG_VALUE); + + /* For floating-point comparisons, try the reverse comparison or try + changing the "orderedness" of the comparison. */ + if (GET_MODE_CLASS (mode) == MODE_FLOAT) + { + enum rtx_code first_code; + bool and_them; + + rcode = reverse_condition_maybe_unordered (code); + if (can_compare_p (rcode, mode, ccp_store_flag) + && (code == ORDERED || code == UNORDERED + || (! HONOR_NANS (mode) && (code == LTGT || code == UNEQ)) + || (! HONOR_SNANS (mode) && (code == EQ || code == NE)))) + { + int want_add = ((STORE_FLAG_VALUE == 1 && normalizep == -1) + || (STORE_FLAG_VALUE == -1 && normalizep == 1)); + + /* For the reverse comparison, use either an addition or a XOR. */ + if (want_add + && rtx_cost (GEN_INT (normalizep), mode, PLUS, 1, + optimize_insn_for_speed_p ()) == 0) + { + tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, + STORE_FLAG_VALUE, target_mode); + if (tem) + return expand_binop (target_mode, add_optab, tem, + gen_int_mode (normalizep, target_mode), + target, 0, OPTAB_WIDEN); + } + else if (!want_add + && rtx_cost (trueval, mode, XOR, 1, + optimize_insn_for_speed_p ()) == 0) + { + tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, + normalizep, target_mode); + if (tem) + return expand_binop (target_mode, xor_optab, tem, trueval, + target, INTVAL (trueval) >= 0, + OPTAB_WIDEN); + } + } + + delete_insns_since (last); + + /* Cannot split ORDERED and UNORDERED, only try the above trick. */ + if (code == ORDERED || code == UNORDERED) + return 0; + + and_them = split_comparison (code, mode, &first_code, &code); + + /* If there are no NaNs, the first comparison should always fall through. + Effectively change the comparison to the other one. */ + if (!HONOR_NANS (mode)) + { + gcc_assert (first_code == (and_them ? ORDERED : UNORDERED)); + return emit_store_flag_1 (target, code, op0, op1, mode, 0, normalizep, + target_mode); + } + + if (!HAVE_conditional_move) + return 0; + + /* Try using a setcc instruction for ORDERED/UNORDERED, followed by a + conditional move. */ + tem = emit_store_flag_1 (subtarget, first_code, op0, op1, mode, 0, + normalizep, target_mode); + if (tem == 0) + return 0; + + if (and_them) + tem = emit_conditional_move (target, code, op0, op1, mode, + tem, const0_rtx, GET_MODE (tem), 0); + else + tem = emit_conditional_move (target, code, op0, op1, mode, + trueval, tem, GET_MODE (tem), 0); + + if (tem == 0) + delete_insns_since (last); + return tem; + } + + /* The remaining tricks only apply to integer comparisons. */ + + if (GET_MODE_CLASS (mode) == MODE_INT) + return emit_store_flag_int (target, subtarget, code, op0, op1, mode, + unsignedp, normalizep, trueval); + + return 0; +} + /* Like emit_store_flag, but always succeeds. */ rtx -- 2.11.4.GIT