Message ID | CAFFmr-7NZef+QOtv2rzcvu4Sc66sTsikGf_gju_fFgGGwi0m_w@mail.gmail.com |
---|---|
State | New, archived |
Headers |
Return-Path: <gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org> Delivered-To: ouuuleilei@gmail.com Received: by 2002:a5d:4ac7:0:0:0:0:0 with SMTP id y7csp223870wrs; Tue, 4 Oct 2022 10:07:32 -0700 (PDT) X-Google-Smtp-Source: AMsMyM4BuBnDqHb/HLtOvYkUlJnJ4lpecQ9t0knGn0J2UN/yZ5Kfwj8JpVjRhhhcldqyuPKhZwce X-Received: by 2002:a05:6402:4150:b0:44a:ec16:def4 with SMTP id x16-20020a056402415000b0044aec16def4mr24702195eda.21.1664903252094; Tue, 04 Oct 2022 10:07:32 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1664903252; cv=none; d=google.com; s=arc-20160816; b=hoYjlEmKuQF4sEsPVqsWNGWHAL0S3TUNfEIFyNvwCS605HRWa/ZyOFk1PuJ3dwujlB 4+KwvFUz8bD1OFU7dsyzpd/wP0IT67ixC6VlQ6HYvWtIgVwi6CuL96zq3tzYPNewEu/1 JEeP0H4pG9miJ37d+aZkqW83pmXkIAnbZYbRbEr/NYulrL0CoWSvY59JjZlP882LlT1Z biG0C2q8ATK1tB0+OSxhD4+L7W8A6S9I4uzz+dKC/axoSyyxVdPv/6UhAqztp9FYp6c/ Id2vssM7vLSYtWdb5kz2MeEBRHSLo2ob3q5irHS28jXFsAuoEIoc4SbamDYUCQZ3FPH+ SK8Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:reply-to:from:list-subscribe:list-help:list-post :list-archive:list-unsubscribe:list-id:precedence:to:subject :message-id:date:mime-version:dmarc-filter:delivered-to :dkim-signature:dkim-filter; bh=pKJXSwPx/V0wP2o6qY0MpKG5FqeWd3wjPD1S8cF67to=; b=yPvtts+mZHUrXfmh/Hrm85BvVsO84qyR2cISgAsH4hvLh3ZPL73ZI74yu14f076Rk7 5BmmwaCF86ddIP77A3xQg2rg+lbR9H3bHwXioT+CkHIJBvxGvKLmI0KWsiXRSevOmDGo MctPu12lGdLRZg59xnAvo4gDHfFNwZtWtJXjOKjHIz1ETQq1s9038FT4fJwlbwVgT7vm iCZMzz5g3J3sSRYKUXqDe9nA6tbSFiHAYZo4FyyLNU83kK35g4KhaRCfWUvyHHeE28NL XHz2u+rtwmeC5V3RNuDN/RJrWwUwEuv7+meHZwlDIaQjqilKfgzuI0paqS5uh935NWHC JJ3A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gcc.gnu.org header.s=default header.b=HVEbkbBZ; spf=pass (google.com: domain of gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org designates 2620:52:3:1:0:246e:9693:128c as permitted sender) smtp.mailfrom="gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=gnu.org Received: from sourceware.org (server2.sourceware.org. [2620:52:3:1:0:246e:9693:128c]) by mx.google.com with ESMTPS id cn28-20020a0564020cbc00b0044de650cc36si10446403edb.147.2022.10.04.10.07.31 for <ouuuleilei@gmail.com> (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 04 Oct 2022 10:07:32 -0700 (PDT) Received-SPF: pass (google.com: domain of gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org designates 2620:52:3:1:0:246e:9693:128c as permitted sender) client-ip=2620:52:3:1:0:246e:9693:128c; Authentication-Results: mx.google.com; dkim=pass header.i=@gcc.gnu.org header.s=default header.b=HVEbkbBZ; spf=pass (google.com: domain of gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org designates 2620:52:3:1:0:246e:9693:128c as permitted sender) smtp.mailfrom="gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=gnu.org Received: from server2.sourceware.org (localhost [IPv6:::1]) by sourceware.org (Postfix) with ESMTP id A05B93858418 for <ouuuleilei@gmail.com>; Tue, 4 Oct 2022 17:07:30 +0000 (GMT) DKIM-Filter: OpenDKIM Filter v2.11.0 sourceware.org A05B93858418 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gcc.gnu.org; s=default; t=1664903250; bh=pKJXSwPx/V0wP2o6qY0MpKG5FqeWd3wjPD1S8cF67to=; h=Date:Subject:To:List-Id:List-Unsubscribe:List-Archive:List-Post: List-Help:List-Subscribe:From:Reply-To:From; b=HVEbkbBZJzU29xNnpZqj2h3FeqnRELVEr9fUJTqsyq3IcWGoFxgd2hS1GmKVh7CWP tbGRcLEOs9cfe5CqznGW2S3P0DUvj04wxQAsQtKoNW05ZniQrS9pVZQ4z3ByVX7XnW 1S7gJeH/jwlTsrZ5SZDDe6314OhDPg/uF040T8qY= X-Original-To: gcc-patches@gcc.gnu.org Delivered-To: gcc-patches@gcc.gnu.org Received: from mail-pl1-x62f.google.com (mail-pl1-x62f.google.com [IPv6:2607:f8b0:4864:20::62f]) by sourceware.org (Postfix) with ESMTPS id 4FFAE3858D32 for <gcc-patches@gcc.gnu.org>; Tue, 4 Oct 2022 17:06:38 +0000 (GMT) DMARC-Filter: OpenDMARC Filter v1.4.1 sourceware.org 4FFAE3858D32 Received: by mail-pl1-x62f.google.com with SMTP id x1so13200005plv.5 for <gcc-patches@gcc.gnu.org>; Tue, 04 Oct 2022 10:06:38 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=to:subject:message-id:date:from:mime-version:x-gm-message-state :from:to:cc:subject:date; bh=Nuxgv+eBsEnfZYSLaR9ltXd4v+iTln1Z6sd9LKakOBQ=; b=V7sCPJXAbv8pfJR3QCOnMKbj0j0kOhFVfhAnf9PbmL8WUpiDhlPn00Vtf+7vtmj5d4 JbiUkBtHI02jui62XTmtFuiXGxAo4XdN1qhmyqT/DoLncg+w/4OFX6QPrtPMznYX4vd0 1ilTerNHriYVBZZwFSI+i78rfMu1YIxVvAQSRXOUTyPWx79qjfFan+ZpAKR0s/P9ZHdE wyCer4a1GAbVj7Dw+nOxZxark1GGpAm7+r4586nehT6WxcpTrQhMh7eUKWqC5814HG7s bFZ34gSnG/jW4GXvLe+BKubarBdEmKAlgJ/PtAcD33Vd+hoQXPpYuPDTcW26pWlGV8/K 2bgw== X-Gm-Message-State: ACrzQf0xHbptpzyqvWnr4Ilt93Hl9QsEp6yqlIaSaaynZmD5Yoz0kr6r KCTgLkzieU8SR0ZRFzq6ZWmCFW/3KR64NCP1+ZIbxM0vl7U= X-Received: by 2002:a17:902:7294:b0:178:a2be:ac18 with SMTP id d20-20020a170902729400b00178a2beac18mr28187556pll.93.1664903195357; Tue, 04 Oct 2022 10:06:35 -0700 (PDT) MIME-Version: 1.0 Date: Tue, 4 Oct 2022 19:06:21 +0200 Message-ID: <CAFFmr-7NZef+QOtv2rzcvu4Sc66sTsikGf_gju_fFgGGwi0m_w@mail.gmail.com> Subject: [PATCH] improved const shifts for AVR targets To: gcc-patches@gcc.gnu.org X-Spam-Status: No, score=-10.2 required=5.0 tests=BAYES_00, DKIM_SIGNED, DKIM_VALID, DKIM_VALID_AU, DKIM_VALID_EF, FREEMAIL_FROM, GIT_PATCH_0, HTML_MESSAGE, RCVD_IN_DNSWL_NONE, SPF_HELO_NONE, SPF_PASS, TXREP autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on server2.sourceware.org Content-Type: text/plain; charset="UTF-8" X-Content-Filtered-By: Mailman/MimeDel 2.1.29 X-BeenThere: gcc-patches@gcc.gnu.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Gcc-patches mailing list <gcc-patches.gcc.gnu.org> List-Unsubscribe: <https://gcc.gnu.org/mailman/options/gcc-patches>, <mailto:gcc-patches-request@gcc.gnu.org?subject=unsubscribe> List-Archive: <https://gcc.gnu.org/pipermail/gcc-patches/> List-Post: <mailto:gcc-patches@gcc.gnu.org> List-Help: <mailto:gcc-patches-request@gcc.gnu.org?subject=help> List-Subscribe: <https://gcc.gnu.org/mailman/listinfo/gcc-patches>, <mailto:gcc-patches-request@gcc.gnu.org?subject=subscribe> From: Alexander Binzberger via Gcc-patches <gcc-patches@gcc.gnu.org> Reply-To: Alexander Binzberger <alexander.binzberger@gmail.com> Errors-To: gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org Sender: "Gcc-patches" <gcc-patches-bounces+ouuuleilei=gmail.com@gcc.gnu.org> X-getmail-retrieved-from-mailbox: =?utf-8?q?INBOX?= X-GMAIL-THRID: =?utf-8?q?1745777592530711507?= X-GMAIL-MSGID: =?utf-8?q?1745777592530711507?= |
Series |
improved const shifts for AVR targets
|
|
Commit Message
A. Binzberger
Oct. 4, 2022, 5:06 p.m. UTC
Hi, recently I used some arduino uno for a project and realized some areas which do not output optimal asm code. Especially around shifts and function calls. With this as motivation and hacktoberfest I started patching things. Since patch files do not provide a good overview and I hope for a "hacktoberfest-accepted" label on the PR on github I also opened it there: https://github.com/gcc-mirror/gcc/pull/73 This patch improves shifts with const right hand operand. While 8bit and 16bit shifts where mostly fine 24bit and 32bit where not handled well. Testing I checked output with a local installation of compiler explorer in asm and a tiny unit test comparing shifts with mul/div by 2. I however did not write any testcases in gcc for it. Target This patch is only targeting atmel avr family of chips. Changelog improved const shifts for AVR targets Patch ----- /* Operand 3 is a scratch register if this is a @@ -6150,96 +6147,68 @@ out_shift_with_cnt (const char *templ, rtx_insn *insn, rtx operands[], /* 8bit shift left ((char)x << i) */ const char * -ashlqi3_out (rtx_insn *insn, rtx operands[], int *len) +ashlqi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { - int k; - - if (!len) - len = &k; - switch (INTVAL (operands[2])) { default: if (INTVAL (operands[2]) < 8) break; - *len = 1; - return "clr %0"; - - case 1: - *len = 1; - return "lsl %0"; - - case 2: - *len = 2; - return ("lsl %0" CR_TAB - "lsl %0"); - - case 3: - *len = 3; - return ("lsl %0" CR_TAB - "lsl %0" CR_TAB - "lsl %0"); + return avr_asm_len ("clr %0", operands, plen, 1); case 4: if (test_hard_reg_class (LD_REGS, operands[0])) { - *len = 2; - return ("swap %0" CR_TAB - "andi %0,0xf0"); + return avr_asm_len ("swap %0" CR_TAB + "andi %0,0xf0", operands, plen, 2); } - *len = 4; - return ("lsl %0" CR_TAB + return avr_asm_len ("lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB - "lsl %0"); + "lsl %0", operands, plen, 4); case 5: if (test_hard_reg_class (LD_REGS, operands[0])) { - *len = 3; - return ("swap %0" CR_TAB + return avr_asm_len ("swap %0" CR_TAB "lsl %0" CR_TAB - "andi %0,0xe0"); + "andi %0,0xe0", operands, plen, 3); } - *len = 5; - return ("lsl %0" CR_TAB + return avr_asm_len ("lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB - "lsl %0"); + "lsl %0", operands, plen, 5); case 6: if (test_hard_reg_class (LD_REGS, operands[0])) { - *len = 4; - return ("swap %0" CR_TAB + return avr_asm_len ("swap %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB - "andi %0,0xc0"); + "andi %0,0xc0", operands, plen, 4); } - *len = 6; - return ("lsl %0" CR_TAB + return avr_asm_len ("lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB "lsl %0" CR_TAB - "lsl %0"); + "lsl %0", operands, plen, 6); case 7: - *len = 3; - return ("ror %0" CR_TAB + return avr_asm_len ("ror %0" CR_TAB "clr %0" CR_TAB - "ror %0"); + "ror %0", operands, plen, 3); } } else if (CONSTANT_P (operands[2])) fatal_insn ("internal compiler error. Incorrect shift:", insn); out_shift_with_cnt ("lsl %0", - insn, operands, len, 1); + insn, operands, plen, 1); return ""; } @@ -6247,7 +6216,7 @@ ashlqi3_out (rtx_insn *insn, rtx operands[], int *len) /* 16bit shift left ((short)x << i) */ const char * -ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) +ashlhi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { @@ -6255,11 +6224,6 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) && XVECLEN (PATTERN (insn), 0) == 3 && REG_P (operands[3])); int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); - int k; - int *t = len; - - if (!len) - len = &k; switch (INTVAL (operands[2])) { @@ -6267,33 +6231,30 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) if (INTVAL (operands[2]) < 16) break; - *len = 2; - return ("clr %B0" CR_TAB - "clr %A0"); + return avr_asm_len ("clr %B0" CR_TAB + "clr %A0", operands, plen, 2); case 4: if (optimize_size && scratch) break; /* 5 */ if (ldi_ok) { - *len = 6; - return ("swap %A0" CR_TAB + return avr_asm_len ("swap %A0" CR_TAB "swap %B0" CR_TAB "andi %B0,0xf0" CR_TAB "eor %B0,%A0" CR_TAB "andi %A0,0xf0" CR_TAB - "eor %B0,%A0"); + "eor %B0,%A0", operands, plen, 6); } if (scratch) { - *len = 7; - return ("swap %A0" CR_TAB + return avr_asm_len ("swap %A0" CR_TAB "swap %B0" CR_TAB "ldi %3,0xf0" CR_TAB "and %B0,%3" CR_TAB "eor %B0,%A0" CR_TAB "and %A0,%3" CR_TAB - "eor %B0,%A0"); + "eor %B0,%A0", operands, plen, 7); } break; /* optimize_size ? 6 : 8 */ @@ -6302,20 +6263,18 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) break; /* scratch ? 5 : 6 */ if (ldi_ok) { - *len = 8; - return ("lsl %A0" CR_TAB + return avr_asm_len ("lsl %A0" CR_TAB "rol %B0" CR_TAB "swap %A0" CR_TAB "swap %B0" CR_TAB "andi %B0,0xf0" CR_TAB "eor %B0,%A0" CR_TAB "andi %A0,0xf0" CR_TAB - "eor %B0,%A0"); + "eor %B0,%A0", operands, plen, 8); } if (scratch) { - *len = 9; - return ("lsl %A0" CR_TAB + return avr_asm_len ("lsl %A0" CR_TAB "rol %B0" CR_TAB "swap %A0" CR_TAB "swap %B0" CR_TAB @@ -6323,15 +6282,14 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) "and %B0,%3" CR_TAB "eor %B0,%A0" CR_TAB "and %A0,%3" CR_TAB - "eor %B0,%A0"); + "eor %B0,%A0", operands, plen, 9); } break; /* 10 */ case 6: if (optimize_size) break; /* scratch ? 5 : 6 */ - *len = 9; - return ("clr __tmp_reg__" CR_TAB + return avr_asm_len ("clr __tmp_reg__" CR_TAB "lsr %B0" CR_TAB "ror %A0" CR_TAB "ror __tmp_reg__" CR_TAB @@ -6339,166 +6297,148 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) "ror %A0" CR_TAB "ror __tmp_reg__" CR_TAB "mov %B0,%A0" CR_TAB - "mov %A0,__tmp_reg__"); + "mov %A0,__tmp_reg__", operands, plen, 9); case 7: - *len = 5; - return ("lsr %B0" CR_TAB + return avr_asm_len ("lsr %B0" CR_TAB "mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "ror %B0" CR_TAB - "ror %A0"); + "ror %A0", operands, plen, 5); case 8: - return *len = 2, ("mov %B0,%A1" CR_TAB - "clr %A0"); + return avr_asm_len ("mov %B0,%A1" CR_TAB + "clr %A0", operands, plen, 2); case 9: - *len = 3; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB - "lsl %B0"); + "lsl %B0", operands, plen, 3); case 10: - *len = 4; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "lsl %B0" CR_TAB - "lsl %B0"); + "lsl %B0", operands, plen, 4); case 11: - *len = 5; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB - "lsl %B0"); + "lsl %B0", operands, plen, 5); case 12: if (ldi_ok) { - *len = 4; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "swap %B0" CR_TAB - "andi %B0,0xf0"); + "andi %B0,0xf0", operands, plen, 4); } if (scratch) { - *len = 5; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "swap %B0" CR_TAB "ldi %3,0xf0" CR_TAB - "and %B0,%3"); + "and %B0,%3", operands, plen, 5); } - *len = 6; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB - "lsl %B0"); + "lsl %B0", operands, plen, 6); case 13: if (ldi_ok) { - *len = 5; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "swap %B0" CR_TAB "lsl %B0" CR_TAB - "andi %B0,0xe0"); + "andi %B0,0xe0", operands, plen, 5); } if (AVR_HAVE_MUL && scratch) { - *len = 5; - return ("ldi %3,0x20" CR_TAB + return avr_asm_len ("ldi %3,0x20" CR_TAB "mul %A0,%3" CR_TAB "mov %B0,r0" CR_TAB "clr %A0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ if (scratch) { - *len = 6; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "swap %B0" CR_TAB "lsl %B0" CR_TAB "ldi %3,0xe0" CR_TAB - "and %B0,%3"); + "and %B0,%3", operands, plen, 6); } if (AVR_HAVE_MUL) { - *len = 6; - return ("set" CR_TAB + return avr_asm_len ("set" CR_TAB "bld r1,5" CR_TAB "mul %A0,r1" CR_TAB "mov %B0,r0" CR_TAB "clr %A0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 6); } - *len = 7; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "clr %A0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB "lsl %B0" CR_TAB - "lsl %B0"); + "lsl %B0", operands, plen, 7); case 14: if (AVR_HAVE_MUL && ldi_ok) { - *len = 5; - return ("ldi %B0,0x40" CR_TAB + return avr_asm_len ("ldi %B0,0x40" CR_TAB "mul %A0,%B0" CR_TAB "mov %B0,r0" CR_TAB "clr %A0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (AVR_HAVE_MUL && scratch) { - *len = 5; - return ("ldi %3,0x40" CR_TAB + return avr_asm_len ("ldi %3,0x40" CR_TAB "mul %A0,%3" CR_TAB "mov %B0,r0" CR_TAB "clr %A0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && ldi_ok) { - *len = 5; - return ("mov %B0,%A0" CR_TAB + return avr_asm_len ("mov %B0,%A0" CR_TAB "ldi %A0,6" "\n1:\t" "lsl %B0" CR_TAB "dec %A0" CR_TAB - "brne 1b"); + "brne 1b", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ - *len = 6; - return ("clr %B0" CR_TAB + return avr_asm_len ("clr %B0" CR_TAB "lsr %A0" CR_TAB "ror %B0" CR_TAB "lsr %A0" CR_TAB "ror %B0" CR_TAB - "clr %A0"); + "clr %A0", operands, plen, 6); case 15: - *len = 4; - return ("clr %B0" CR_TAB + return avr_asm_len ("clr %B0" CR_TAB "lsr %A0" CR_TAB "ror %B0" CR_TAB - "clr %A0"); + "clr %A0", operands, plen, 4); } - len = t; } out_shift_with_cnt ("lsl %A0" CR_TAB - "rol %B0", insn, operands, len, 2); + "rol %B0", insn, operands, plen, 2); return ""; } @@ -6506,13 +6446,22 @@ ashlhi3_out (rtx_insn *insn, rtx operands[], int *len) /* 24-bit shift left */ const char* -avr_out_ashlpsi3 (rtx_insn *insn, rtx *op, int *plen) +avr_out_ashlpsi3 (rtx_insn *insn, rtx *operands, int *plen) { - if (plen) - *plen = 0; + rtx op[5]; + + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; if (CONST_INT_P (op[2])) { + int scratch = (GET_CODE (PATTERN (insn)) == PARALLEL + && XVECLEN (PATTERN (insn), 0) == 3 + && REG_P (operands[3])); + int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); + switch (INTVAL (op[2])) { default: @@ -6523,31 +6472,145 @@ avr_out_ashlpsi3 (rtx_insn *insn, rtx *op, int *plen) "clr %B0" CR_TAB "clr %C0", op, plen, 3); + case 4: + if (optimize_size && scratch) + break; /* 6 */ + if (ldi_ok) + { + return avr_asm_len ("swap %C0" CR_TAB + "andi %C0, 0xf0" CR_TAB + "swap %B0" CR_TAB + "eor %C0, %B0" CR_TAB + "andi %B0, 0xf0" CR_TAB + "eor %C0, %B0" CR_TAB + "swap %A0" CR_TAB + "eor %B0, %A0" CR_TAB + "andi %A0, 0xf0" CR_TAB + "eor %B0, %A0", op, plen, 10); + } + if (scratch) + { + return avr_asm_len ("swap %C0" CR_TAB + "ldi %3,0xf0" CR_TAB + "and %C0, %3" CR_TAB + "swap %B0" CR_TAB + "eor %C0, %B0" CR_TAB + "and %B0, %3" CR_TAB + "eor %C0, %B0" CR_TAB + "swap %A0" CR_TAB + "eor %B0, %A0" CR_TAB + "and %A0, %3" CR_TAB + "eor %B0, %A0", op, plen, 11); + } + break; /* optimize_size ? 7 : 9 */ + + case 5: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + if (ldi_ok) + { + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "swap %C0" CR_TAB + "andi %C0, 0xf0" CR_TAB + "swap %B0" CR_TAB + "eor %C0, %B0" CR_TAB + "andi %B0, 0xf0" CR_TAB + "eor %C0, %B0" CR_TAB + "swap %A0" CR_TAB + "eor %B0, %A0" CR_TAB + "andi %A0, 0xf0" CR_TAB + "eor %B0, %A0", op, plen, 13); + } + if (scratch) + { + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "swap %C0" CR_TAB + "ldi %3,0xf0" CR_TAB + "and %C0, %3" CR_TAB + "swap %B0" CR_TAB + "eor %C0, %B0" CR_TAB + "and %B0, %3" CR_TAB + "eor %C0, %B0" CR_TAB + "swap %A0" CR_TAB + "eor %B0, %A0" CR_TAB + "and %A0, %3" CR_TAB + "eor %B0, %A0", op, plen, 14); + } + break; /* 10 */ + + case 6: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + return avr_asm_len ("clr __tmp_reg__" CR_TAB + "lsr %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "ror __tmp_reg__" CR_TAB + "lsr %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "ror __tmp_reg__" CR_TAB + "mov %C0,%B0" CR_TAB + "mov %B0,%A0" CR_TAB + "mov %A0,__tmp_reg__", op, plen, 12); + + case 7: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + return avr_asm_len ("lsr %C0" CR_TAB + "mov %C0,%B0" CR_TAB + "mov %B0,%A0" CR_TAB + "clr %A0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0", op, plen, 7); + case 8: + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: { int reg0 = REGNO (op[0]); int reg1 = REGNO (op[1]); if (reg0 >= reg1) - return avr_asm_len ("mov %C0,%B1" CR_TAB - "mov %B0,%A1" CR_TAB - "clr %A0", op, plen, 3); + avr_asm_len("mov %C0,%B1" CR_TAB + "mov %B0,%A1" CR_TAB + "clr %A0", op, plen, 3); else - return avr_asm_len ("clr %A0" CR_TAB - "mov %B0,%A1" CR_TAB - "mov %C0,%B1", op, plen, 3); + avr_asm_len("clr %A0" CR_TAB + "mov %B0,%A1" CR_TAB + "mov %C0,%B1", op, plen, 3); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return ashlhi3_out(insn, op, plen); } case 16: + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: { int reg0 = REGNO (op[0]); int reg1 = REGNO (op[1]); if (reg0 + 2 != reg1) - avr_asm_len ("mov %C0,%A0", op, plen, 1); + avr_asm_len("mov %C0,%A0", op, plen, 1); - return avr_asm_len ("clr %B0" CR_TAB - "clr %A0", op, plen, 2); + avr_asm_len("clr %B0" CR_TAB + "clr %A0", op, plen, 2); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return ashlqi3_out(insn, op, plen); } case 23: @@ -6569,15 +6632,21 @@ avr_out_ashlpsi3 (rtx_insn *insn, rtx *op, int *plen) /* 32bit shift left ((long)x << i) */ const char * -ashlsi3_out (rtx_insn *insn, rtx operands[], int *len) +ashlsi3_out (rtx_insn *insn, rtx operands[], int *plen) { + rtx op[5]; + + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; + if (CONST_INT_P (operands[2])) { - int k; - int *t = len; - - if (!len) - len = &k; + int scratch = (GET_CODE (PATTERN (insn)) == PARALLEL + && XVECLEN (PATTERN (insn), 0) == 3 + && REG_P (operands[3])); + int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); switch (INTVAL (operands[2])) { @@ -6586,125 +6655,247 @@ ashlsi3_out (rtx_insn *insn, rtx operands[], int *len) break; if (AVR_HAVE_MOVW) - return *len = 3, ("clr %D0" CR_TAB + return avr_asm_len ("clr %D0" CR_TAB "clr %C0" CR_TAB - "movw %A0,%C0"); - *len = 4; - return ("clr %D0" CR_TAB + "movw %A0,%C0", operands, plen, 3); + + return avr_asm_len ("clr %D0" CR_TAB "clr %C0" CR_TAB "clr %B0" CR_TAB - "clr %A0"); + "clr %A0", operands, plen, 4); - case 8: - { + case 4: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + if (ldi_ok) + { + return avr_asm_len ("swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %D0, 0xf0" CR_TAB + "eor %D0, %C0" CR_TAB + "andi %C0, 0xf0" CR_TAB + "eor %D0, %C0" CR_TAB + "eor %C0, %B0" CR_TAB + "andi %B0, 0xf0" CR_TAB + "eor %C0, %B0" CR_TAB + "eor %B0, %A0" CR_TAB + "andi %A0, 0xf0" CR_TAB + "eor %B0, %A0", op, plen, 14); + } + if (scratch) + { + return avr_asm_len ("swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0xf0" CR_TAB + "and %D0, %3" CR_TAB + "eor %D0, %C0" CR_TAB + "and %C0, %3" CR_TAB + "eor %D0, %C0" CR_TAB + "eor %C0, %B0" CR_TAB + "and %B0, %3" CR_TAB + "eor %C0, %B0" CR_TAB + "eor %B0, %A0" CR_TAB + "and %A0, %3" CR_TAB + "eor %B0, %A0", op, plen, 15); + } + break; /* 10 */ + + case 5: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + if (ldi_ok) + { + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %D0, 0xf0" CR_TAB + "eor %D0, %C0" CR_TAB + "andi %C0, 0xf0" CR_TAB + "eor %D0, %C0" CR_TAB + "eor %C0, %B0" CR_TAB + "andi %B0, 0xf0" CR_TAB + "eor %C0, %B0" CR_TAB + "eor %B0, %A0" CR_TAB + "andi %A0, 0xf0" CR_TAB + "eor %B0, %A0", op, plen, 18); + } + if (scratch) + { + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0xf0" CR_TAB + "and %D0, %3" CR_TAB + "eor %D0, %C0" CR_TAB + "and %C0, %3" CR_TAB + "eor %D0, %C0" CR_TAB + "eor %C0, %B0" CR_TAB + "and %B0, %3" CR_TAB + "eor %C0, %B0" CR_TAB + "eor %B0, %A0" CR_TAB + "and %A0, %3" CR_TAB + "eor %B0, %A0", op, plen, 19); + } + break; /* 10 */ + + case 6: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + return avr_asm_len ("clr __tmp_reg__" CR_TAB + "lsr %D0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "ror __tmp_reg__" CR_TAB + "lsr %D0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "ror __tmp_reg__" CR_TAB + "mov %D0,%C0" CR_TAB + "mov %C0,%B0" CR_TAB + "mov %B0,%A0" CR_TAB + "mov %A0,__tmp_reg__", op, plen, 15); + + case 7: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + return avr_asm_len ("lsr %D0" CR_TAB + "mov %D0,%C0" CR_TAB + "mov %C0,%B0" CR_TAB + "mov %B0,%A0" CR_TAB + "clr %A0" CR_TAB + "ror %D0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0", op, plen, 9); + + case 8: + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); - *len = 4; if (reg0 >= reg1) - return ("mov %D0,%C1" CR_TAB - "mov %C0,%B1" CR_TAB - "mov %B0,%A1" CR_TAB - "clr %A0"); + avr_asm_len("mov %D0,%C1" CR_TAB + "mov %C0,%B1" CR_TAB + "mov %B0,%A1" CR_TAB + "clr %A0", operands, plen, 4); else - return ("clr %A0" CR_TAB - "mov %B0,%A1" CR_TAB - "mov %C0,%B1" CR_TAB - "mov %D0,%C1"); - } - - case 16: - { + avr_asm_len("clr %A0" CR_TAB + "mov %B0,%A1" CR_TAB + "mov %C0,%B1" CR_TAB + "mov %D0,%C1", operands, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return avr_out_ashlpsi3(insn, op, plen); + } + + case 16: + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: + case 23: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); if (reg0 + 2 == reg1) - return *len = 2, ("clr %B0" CR_TAB - "clr %A0"); + avr_asm_len("clr %B0" CR_TAB + "clr %A0", op, plen, 2); if (AVR_HAVE_MOVW) - return *len = 3, ("movw %C0,%A1" CR_TAB - "clr %B0" CR_TAB - "clr %A0"); + avr_asm_len("movw %C0,%A1" CR_TAB + "clr %B0" CR_TAB + "clr %A0", op, plen, 3); else - return *len = 4, ("mov %C0,%A1" CR_TAB - "mov %D0,%B1" CR_TAB - "clr %B0" CR_TAB - "clr %A0"); - } - - case 24: - *len = 4; - return ("mov %D0,%A1" CR_TAB - "clr %C0" CR_TAB - "clr %B0" CR_TAB - "clr %A0"); - - case 31: - *len = 6; - return ("clr %D0" CR_TAB + avr_asm_len("mov %C0,%A1" CR_TAB + "mov %D0,%B1" CR_TAB + "clr %B0" CR_TAB + "clr %A0", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return ashlhi3_out(insn, op, plen); + } + + case 24: + case 25: + case 26: + case 27: + case 28: + case 29: + case 30: + { + avr_asm_len("mov %D0,%A1" CR_TAB + "clr %C0" CR_TAB + "clr %B0" CR_TAB + "clr %A0", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 24); + return ashlqi3_out(insn, op, plen); + } + + case 31: + return avr_asm_len ("clr %D0" CR_TAB "lsr %A0" CR_TAB "ror %D0" CR_TAB "clr %C0" CR_TAB "clr %B0" CR_TAB - "clr %A0"); + "clr %A0", operands, plen, 6); } - len = t; } out_shift_with_cnt ("lsl %A0" CR_TAB "rol %B0" CR_TAB "rol %C0" CR_TAB - "rol %D0", insn, operands, len, 4); + "rol %D0", insn, op, plen, 4); return ""; } /* 8bit arithmetic shift right ((signed char)x >> i) */ const char * -ashrqi3_out (rtx_insn *insn, rtx operands[], int *len) +ashrqi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { - int k; - - if (!len) - len = &k; - switch (INTVAL (operands[2])) { - case 1: - *len = 1; - return "asr %0"; - - case 2: - *len = 2; - return ("asr %0" CR_TAB - "asr %0"); - case 3: - *len = 3; - return ("asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0"); - - case 4: - *len = 4; - return ("asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0"); - - case 5: - *len = 5; - return ("asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0" CR_TAB - "asr %0"); - - case 6: - *len = 4; - return ("bst %0,6" CR_TAB + case 4: + return avr_asm_len ("asr %0" CR_TAB + "asr %0" CR_TAB + "asr %0" CR_TAB + "asr %0", operands, plen, 4); + + case 5: + return avr_asm_len ("asr %0" CR_TAB + "asr %0" CR_TAB + "asr %0" CR_TAB + "asr %0" CR_TAB + "asr %0", operands, plen, 5); + + case 6: + return avr_asm_len ("bst %0,6" CR_TAB "lsl %0" CR_TAB "sbc %0,%0" CR_TAB - "bld %0,0"); + "bld %0,0", operands, plen, 4); default: if (INTVAL (operands[2]) < 8) @@ -6713,16 +6904,15 @@ ashrqi3_out (rtx_insn *insn, rtx operands[], int *len) /* fall through */ case 7: - *len = 2; - return ("lsl %0" CR_TAB - "sbc %0,%0"); + return avr_asm_len ("lsl %0" CR_TAB + "sbc %0,%0", operands, plen, 2); } } else if (CONSTANT_P (operands[2])) fatal_insn ("internal compiler error. Incorrect shift:", insn); out_shift_with_cnt ("asr %0", - insn, operands, len, 1); + insn, operands, plen, 1); return ""; } @@ -6730,7 +6920,7 @@ ashrqi3_out (rtx_insn *insn, rtx operands[], int *len) /* 16bit arithmetic shift right ((signed short)x >> i) */ const char * -ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) +ashrhi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { @@ -6738,11 +6928,6 @@ ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) && XVECLEN (PATTERN (insn), 0) == 3 && REG_P (operands[3])); int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); - int k; - int *t = len; - - if (!len) - len = &k; switch (INTVAL (operands[2])) { @@ -6754,22 +6939,20 @@ ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) case 6: if (optimize_size) break; /* scratch ? 5 : 6 */ - *len = 8; - return ("mov __tmp_reg__,%A0" CR_TAB + return avr_asm_len ("mov __tmp_reg__,%A0" CR_TAB "mov %A0,%B0" CR_TAB "lsl __tmp_reg__" CR_TAB "rol %A0" CR_TAB "sbc %B0,%B0" CR_TAB "lsl __tmp_reg__" CR_TAB "rol %A0" CR_TAB - "rol %B0"); + "rol %B0", operands, plen, 8); case 7: - *len = 4; - return ("lsl %A0" CR_TAB + return avr_asm_len ("lsl %A0" CR_TAB "mov %A0,%B0" CR_TAB "rol %A0" CR_TAB - "sbc %B0,%B0"); + "sbc %B0,%B0", operands, plen, 4); case 8: { @@ -6777,101 +6960,92 @@ ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) int reg1 = true_regnum (operands[1]); if (reg0 == reg1) - return *len = 3, ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB - "sbc %B0,%B0"); + "sbc %B0,%B0", operands, plen, 3); else - return *len = 4, ("mov %A0,%B1" CR_TAB + return avr_asm_len ("mov %A0,%B1" CR_TAB "clr %B0" CR_TAB "sbrc %A0,7" CR_TAB - "dec %B0"); + "dec %B0", operands, plen, 4); } case 9: - *len = 4; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB "sbc %B0,%B0" CR_TAB - "asr %A0"); + "asr %A0", operands, plen, 4); case 10: - *len = 5; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB "sbc %B0,%B0" CR_TAB "asr %A0" CR_TAB - "asr %A0"); + "asr %A0", operands, plen, 5); case 11: if (AVR_HAVE_MUL && ldi_ok) { - *len = 5; - return ("ldi %A0,0x20" CR_TAB + return avr_asm_len ("ldi %A0,0x20" CR_TAB "muls %B0,%A0" CR_TAB "mov %A0,r1" CR_TAB "sbc %B0,%B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ - *len = 6; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB "sbc %B0,%B0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB - "asr %A0"); + "asr %A0", operands, plen, 6); case 12: if (AVR_HAVE_MUL && ldi_ok) { - *len = 5; - return ("ldi %A0,0x10" CR_TAB + return avr_asm_len ("ldi %A0,0x10" CR_TAB "muls %B0,%A0" CR_TAB "mov %A0,r1" CR_TAB "sbc %B0,%B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ - *len = 7; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB "sbc %B0,%B0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB - "asr %A0"); + "asr %A0", operands, plen, 7); case 13: if (AVR_HAVE_MUL && ldi_ok) { - *len = 5; - return ("ldi %A0,0x08" CR_TAB + return avr_asm_len ("ldi %A0,0x08" CR_TAB "muls %B0,%A0" CR_TAB "mov %A0,r1" CR_TAB "sbc %B0,%B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size) - break; /* scratch ? 5 : 7 */ - *len = 8; - return ("mov %A0,%B0" CR_TAB + break; /* scratch ? 5 : 6 */ + return avr_asm_len ("mov %A0,%B0" CR_TAB "lsl %B0" CR_TAB "sbc %B0,%B0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB "asr %A0" CR_TAB - "asr %A0"); + "asr %A0", operands, plen, 8); case 14: - *len = 5; - return ("lsl %B0" CR_TAB + return avr_asm_len ("lsl %B0" CR_TAB "sbc %A0,%A0" CR_TAB "lsl %B0" CR_TAB "mov %B0,%A0" CR_TAB - "rol %A0"); + "rol %A0", operands, plen, 5); default: if (INTVAL (operands[2]) < 16) @@ -6880,14 +7054,13 @@ ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) /* fall through */ case 15: - return *len = 3, ("lsl %B0" CR_TAB + return avr_asm_len ("lsl %B0" CR_TAB "sbc %A0,%A0" CR_TAB - "mov %B0,%A0"); + "mov %B0,%A0", operands, plen, 3); } - len = t; } out_shift_with_cnt ("asr %B0" CR_TAB - "ror %A0", insn, operands, len, 2); + "ror %A0", insn, operands, plen, 2); return ""; } @@ -6895,40 +7068,93 @@ ashrhi3_out (rtx_insn *insn, rtx operands[], int *len) /* 24-bit arithmetic shift right */ const char* -avr_out_ashrpsi3 (rtx_insn *insn, rtx *op, int *plen) +avr_out_ashrpsi3 (rtx_insn *insn, rtx *operands, int *plen) { + rtx op[5]; + + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; + int dest = REGNO (op[0]); int src = REGNO (op[1]); if (CONST_INT_P (op[2])) { - if (plen) - *plen = 0; - switch (INTVAL (op[2])) { + case 4: + case 5: + /* XXX try to optimize this too? */ + break; + + case 6: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + return avr_asm_len ("mov __tmp_reg__,%A0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "lsl __tmp_reg__" CR_TAB + "rol %A0" CR_TAB + "rol %B0" CR_TAB + "sbc %C0,%C0" CR_TAB + "lsl __tmp_reg__" CR_TAB + "rol %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0", operands, plen, 11); + + case 7: + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "sbc %C0,%C0", operands, plen, 6); + case 8: - if (dest <= src) - return avr_asm_len ("mov %A0,%B1" CR_TAB - "mov %B0,%C1" CR_TAB - "clr %C0" CR_TAB - "sbrc %B0,7" CR_TAB - "dec %C0", op, plen, 5); - else - return avr_asm_len ("clr %C0" CR_TAB - "sbrc %C1,7" CR_TAB - "dec %C0" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %A0,%B1", op, plen, 5); + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: + { + if (dest <= src) + avr_asm_len("mov %A0,%B1" CR_TAB + "mov %B0,%C1" CR_TAB + "clr %C0" CR_TAB + "sbrc %B0,7" CR_TAB + "dec %C0", op, plen, 5); + else + avr_asm_len("clr %C0" CR_TAB + "sbrc %C1,7" CR_TAB + "dec %C0" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %A0,%B1", op, plen, 5); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return ashrhi3_out(insn, op, plen); + } case 16: - if (dest != src + 2) - avr_asm_len ("mov %A0,%C1", op, plen, 1); - - return avr_asm_len ("clr %B0" CR_TAB - "sbrc %A0,7" CR_TAB - "com %B0" CR_TAB - "mov %C0,%B0", op, plen, 4); + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: + { + if (dest != src + 2) + avr_asm_len("mov %A0,%C1", op, plen, 1); + + avr_asm_len("clr %B0" CR_TAB + "sbrc %A0,7" CR_TAB + "com %B0" CR_TAB + "mov %C0,%B0", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return ashrqi3_out(insn, op, plen); + } default: if (INTVAL (op[2]) < 24) @@ -6954,71 +7180,136 @@ avr_out_ashrpsi3 (rtx_insn *insn, rtx *op, int *plen) /* 32-bit arithmetic shift right ((signed long)x >> i) */ const char * -ashrsi3_out (rtx_insn *insn, rtx operands[], int *len) +ashrsi3_out (rtx_insn *insn, rtx operands[], int *plen) { - if (CONST_INT_P (operands[2])) - { - int k; - int *t = len; + rtx op[5]; - if (!len) - len = &k; + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; + if (CONST_INT_P (operands[2])) + { switch (INTVAL (operands[2])) { - case 8: - { + case 4: + case 5: + /* XXX try to optimize this too? */ + break; + + case 6: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + return avr_asm_len ("mov __tmp_reg__,%A0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "mov %C0,%D0" CR_TAB + "lsl __tmp_reg__" CR_TAB + "rol %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "sbc %D0,%D0" CR_TAB + "lsl __tmp_reg__" CR_TAB + "rol %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0", operands, plen, 14); + + case 7: + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "mov %C0,%D0" CR_TAB + "sbc %D0,%D0", operands, plen, 8); + + case 8: + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); - *len=6; if (reg0 <= reg1) - return ("mov %A0,%B1" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %C0,%D1" CR_TAB - "clr %D0" CR_TAB - "sbrc %C0,7" CR_TAB - "dec %D0"); + avr_asm_len("mov %A0,%B1" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %C0,%D1" CR_TAB + "clr %D0" CR_TAB + "sbrc %C0,7" CR_TAB + "dec %D0", op, plen, 6); else - return ("clr %D0" CR_TAB - "sbrc %D1,7" CR_TAB - "dec %D0" CR_TAB - "mov %C0,%D1" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %A0,%B1"); - } - - case 16: - { + avr_asm_len("clr %D0" CR_TAB + "sbrc %D1,7" CR_TAB + "dec %D0" CR_TAB + "mov %C0,%D1" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %A0,%B1", op, plen, 6); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return avr_out_ashrpsi3(insn, op, plen); + } + + case 16: + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: + case 23: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); if (reg0 == reg1 + 2) - return *len = 4, ("clr %D0" CR_TAB - "sbrc %B0,7" CR_TAB - "com %D0" CR_TAB - "mov %C0,%D0"); + avr_asm_len("clr %D0" CR_TAB + "sbrc %B0,7" CR_TAB + "com %D0" CR_TAB + "mov %C0,%D0", op, plen, 4); if (AVR_HAVE_MOVW) - return *len = 5, ("movw %A0,%C1" CR_TAB - "clr %D0" CR_TAB - "sbrc %B0,7" CR_TAB - "com %D0" CR_TAB - "mov %C0,%D0"); + avr_asm_len("movw %A0,%C1" CR_TAB + "clr %D0" CR_TAB + "sbrc %B0,7" CR_TAB + "com %D0" CR_TAB + "mov %C0,%D0", op, plen, 5); else - return *len = 6, ("mov %B0,%D1" CR_TAB - "mov %A0,%C1" CR_TAB - "clr %D0" CR_TAB - "sbrc %B0,7" CR_TAB - "com %D0" CR_TAB - "mov %C0,%D0"); - } - - case 24: - return *len = 6, ("mov %A0,%D1" CR_TAB + avr_asm_len("mov %B0,%D1" CR_TAB + "mov %A0,%C1" CR_TAB + "clr %D0" CR_TAB + "sbrc %B0,7" CR_TAB + "com %D0" CR_TAB + "mov %C0,%D0", op, plen, 6); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return ashrhi3_out(insn, op, plen); + } + + case 24: + case 25: + case 26: + case 27: + case 28: + case 29: + case 30: + { + int reg0 = true_regnum (operands[0]); + int reg1 = true_regnum (operands[1]); + + if (reg0 == reg1 + 2) + avr_asm_len("mov %A0,%D1" CR_TAB "clr %D0" CR_TAB "sbrc %A0,7" CR_TAB "com %D0" CR_TAB "mov %B0,%D0" CR_TAB - "mov %C0,%D0"); + "mov %C0,%D0", op, plen, 6); + op[2] = GEN_INT((INTVAL (operands[2])) - 24); + return ashrqi3_out(insn, op, plen); + } default: if (INTVAL (operands[2]) < 32) @@ -7028,37 +7319,33 @@ ashrsi3_out (rtx_insn *insn, rtx operands[], int *len) case 31: if (AVR_HAVE_MOVW) - return *len = 4, ("lsl %D0" CR_TAB + return avr_asm_len ("lsl %D0" CR_TAB "sbc %A0,%A0" CR_TAB "mov %B0,%A0" CR_TAB - "movw %C0,%A0"); + "movw %C0,%A0", operands, plen, 4); else - return *len = 5, ("lsl %D0" CR_TAB + return avr_asm_len ("lsl %D0" CR_TAB "sbc %A0,%A0" CR_TAB "mov %B0,%A0" CR_TAB "mov %C0,%A0" CR_TAB - "mov %D0,%A0"); + "mov %D0,%A0", operands, plen, 5); } - len = t; } out_shift_with_cnt ("asr %D0" CR_TAB "ror %C0" CR_TAB "ror %B0" CR_TAB - "ror %A0", insn, operands, len, 4); + "ror %A0", insn, operands, plen, 4); return ""; } /* 8-bit logic shift right ((unsigned char)x >> i) */ const char * -lshrqi3_out (rtx_insn *insn, rtx operands[], int *len) +lshrqi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { - int k; - - if (!len) - len = &k; + int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); switch (INTVAL (operands[2])) { @@ -7066,87 +7353,65 @@ lshrqi3_out (rtx_insn *insn, rtx operands[], int *len) if (INTVAL (operands[2]) < 8) break; - *len = 1; - return "clr %0"; - - case 1: - *len = 1; - return "lsr %0"; - - case 2: - *len = 2; - return ("lsr %0" CR_TAB - "lsr %0"); - case 3: - *len = 3; - return ("lsr %0" CR_TAB - "lsr %0" CR_TAB - "lsr %0"); + return avr_asm_len("clr %0", operands, plen, 1); case 4: - if (test_hard_reg_class (LD_REGS, operands[0])) + if (ldi_ok) { - *len=2; - return ("swap %0" CR_TAB - "andi %0,0x0f"); + return avr_asm_len ("swap %0" CR_TAB + "andi %0,0x0f", operands, plen, 2); } - *len = 4; - return ("lsr %0" CR_TAB + return avr_asm_len ("lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB - "lsr %0"); + "lsr %0", operands, plen, 4); case 5: - if (test_hard_reg_class (LD_REGS, operands[0])) + if (ldi_ok) { - *len = 3; - return ("swap %0" CR_TAB + return avr_asm_len ("swap %0" CR_TAB "lsr %0" CR_TAB - "andi %0,0x7"); + "andi %0,0x7", operands, plen, 3); } - *len = 5; - return ("lsr %0" CR_TAB + return avr_asm_len ("lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB - "lsr %0"); + "lsr %0", operands, plen, 5); case 6: - if (test_hard_reg_class (LD_REGS, operands[0])) + if (ldi_ok) { - *len = 4; - return ("swap %0" CR_TAB + return avr_asm_len ("swap %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB - "andi %0,0x3"); + "andi %0,0x3", operands, plen, 4); } - *len = 6; - return ("lsr %0" CR_TAB + return avr_asm_len ("lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB "lsr %0" CR_TAB - "lsr %0"); + "lsr %0", operands, plen, 6); case 7: - *len = 3; - return ("rol %0" CR_TAB + return avr_asm_len ("rol %0" CR_TAB "clr %0" CR_TAB - "rol %0"); + "rol %0", operands, plen, 3); } } else if (CONSTANT_P (operands[2])) fatal_insn ("internal compiler error. Incorrect shift:", insn); out_shift_with_cnt ("lsr %0", - insn, operands, len, 1); + insn, operands, plen, 1); return ""; } /* 16-bit logic shift right ((unsigned short)x >> i) */ const char * -lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) +lshrhi3_out (rtx_insn *insn, rtx operands[], int *plen) { if (CONST_INT_P (operands[2])) { @@ -7154,11 +7419,6 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) && XVECLEN (PATTERN (insn), 0) == 3 && REG_P (operands[3])); int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); - int k; - int *t = len; - - if (!len) - len = &k; switch (INTVAL (operands[2])) { @@ -7166,33 +7426,30 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) if (INTVAL (operands[2]) < 16) break; - *len = 2; - return ("clr %B0" CR_TAB - "clr %A0"); + return avr_asm_len ("clr %B0" CR_TAB + "clr %A0", operands, plen, 2); case 4: if (optimize_size && scratch) break; /* 5 */ if (ldi_ok) { - *len = 6; - return ("swap %B0" CR_TAB + return avr_asm_len ("swap %B0" CR_TAB "swap %A0" CR_TAB "andi %A0,0x0f" CR_TAB "eor %A0,%B0" CR_TAB "andi %B0,0x0f" CR_TAB - "eor %A0,%B0"); + "eor %A0,%B0", operands, plen, 6); } if (scratch) { - *len = 7; - return ("swap %B0" CR_TAB + return avr_asm_len ("swap %B0" CR_TAB "swap %A0" CR_TAB "ldi %3,0x0f" CR_TAB "and %A0,%3" CR_TAB "eor %A0,%B0" CR_TAB "and %B0,%3" CR_TAB - "eor %A0,%B0"); + "eor %A0,%B0", operands, plen, 7); } break; /* optimize_size ? 6 : 8 */ @@ -7201,20 +7458,18 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) break; /* scratch ? 5 : 6 */ if (ldi_ok) { - *len = 8; - return ("lsr %B0" CR_TAB + return avr_asm_len ("lsr %B0" CR_TAB "ror %A0" CR_TAB "swap %B0" CR_TAB "swap %A0" CR_TAB "andi %A0,0x0f" CR_TAB "eor %A0,%B0" CR_TAB "andi %B0,0x0f" CR_TAB - "eor %A0,%B0"); + "eor %A0,%B0", operands, plen, 8); } if (scratch) { - *len = 9; - return ("lsr %B0" CR_TAB + return avr_asm_len ("lsr %B0" CR_TAB "ror %A0" CR_TAB "swap %B0" CR_TAB "swap %A0" CR_TAB @@ -7222,15 +7477,14 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) "and %A0,%3" CR_TAB "eor %A0,%B0" CR_TAB "and %B0,%3" CR_TAB - "eor %A0,%B0"); + "eor %A0,%B0", operands, plen, 9); } break; /* 10 */ case 6: if (optimize_size) break; /* scratch ? 5 : 6 */ - *len = 9; - return ("clr __tmp_reg__" CR_TAB + return avr_asm_len ("clr __tmp_reg__" CR_TAB "lsl %A0" CR_TAB "rol %B0" CR_TAB "rol __tmp_reg__" CR_TAB @@ -7238,166 +7492,148 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) "rol %B0" CR_TAB "rol __tmp_reg__" CR_TAB "mov %A0,%B0" CR_TAB - "mov %B0,__tmp_reg__"); + "mov %B0,__tmp_reg__", operands, plen, 9); case 7: - *len = 5; - return ("lsl %A0" CR_TAB + return avr_asm_len ("lsl %A0" CR_TAB "mov %A0,%B0" CR_TAB "rol %A0" CR_TAB "sbc %B0,%B0" CR_TAB - "neg %B0"); + "neg %B0", operands, plen, 5); case 8: - return *len = 2, ("mov %A0,%B1" CR_TAB - "clr %B0"); + return avr_asm_len ("mov %A0,%B1" CR_TAB + "clr %B0", operands, plen, 2); case 9: - *len = 3; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB - "lsr %A0"); + "lsr %A0", operands, plen, 3); case 10: - *len = 4; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "lsr %A0" CR_TAB - "lsr %A0"); + "lsr %A0", operands, plen, 4); case 11: - *len = 5; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB - "lsr %A0"); + "lsr %A0", operands, plen, 5); case 12: if (ldi_ok) { - *len = 4; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "swap %A0" CR_TAB - "andi %A0,0x0f"); + "andi %A0,0x0f", operands, plen, 4); } if (scratch) { - *len = 5; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "swap %A0" CR_TAB "ldi %3,0x0f" CR_TAB - "and %A0,%3"); + "and %A0,%3", operands, plen, 5); } - *len = 6; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB - "lsr %A0"); + "lsr %A0", operands, plen, 6); case 13: if (ldi_ok) { - *len = 5; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "swap %A0" CR_TAB "lsr %A0" CR_TAB - "andi %A0,0x07"); + "andi %A0,0x07", operands, plen, 5); } if (AVR_HAVE_MUL && scratch) { - *len = 5; - return ("ldi %3,0x08" CR_TAB + return avr_asm_len ("ldi %3,0x08" CR_TAB "mul %B0,%3" CR_TAB "mov %A0,r1" CR_TAB "clr %B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ if (scratch) { - *len = 6; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "swap %A0" CR_TAB "lsr %A0" CR_TAB "ldi %3,0x07" CR_TAB - "and %A0,%3"); + "and %A0,%3", operands, plen, 6); } if (AVR_HAVE_MUL) { - *len = 6; - return ("set" CR_TAB + return avr_asm_len ("set" CR_TAB "bld r1,3" CR_TAB "mul %B0,r1" CR_TAB "mov %A0,r1" CR_TAB "clr %B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 6); } - *len = 7; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "clr %B0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB "lsr %A0" CR_TAB - "lsr %A0"); + "lsr %A0", operands, plen, 7); case 14: if (AVR_HAVE_MUL && ldi_ok) { - *len = 5; - return ("ldi %A0,0x04" CR_TAB + return avr_asm_len ("ldi %A0,0x04" CR_TAB "mul %B0,%A0" CR_TAB "mov %A0,r1" CR_TAB "clr %B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (AVR_HAVE_MUL && scratch) { - *len = 5; - return ("ldi %3,0x04" CR_TAB + return avr_asm_len ("ldi %3,0x04" CR_TAB "mul %B0,%3" CR_TAB "mov %A0,r1" CR_TAB "clr %B0" CR_TAB - "clr __zero_reg__"); + "clr __zero_reg__", operands, plen, 5); } if (optimize_size && ldi_ok) { - *len = 5; - return ("mov %A0,%B0" CR_TAB + return avr_asm_len ("mov %A0,%B0" CR_TAB "ldi %B0,6" "\n1:\t" "lsr %A0" CR_TAB "dec %B0" CR_TAB - "brne 1b"); + "brne 1b", operands, plen, 5); } if (optimize_size && scratch) break; /* 5 */ - *len = 6; - return ("clr %A0" CR_TAB + return avr_asm_len ("clr %A0" CR_TAB "lsl %B0" CR_TAB "rol %A0" CR_TAB "lsl %B0" CR_TAB "rol %A0" CR_TAB - "clr %B0"); + "clr %B0", operands, plen, 6); case 15: - *len = 4; - return ("clr %A0" CR_TAB + return avr_asm_len ("clr %A0" CR_TAB "lsl %B0" CR_TAB "rol %A0" CR_TAB - "clr %B0"); + "clr %B0", operands, plen, 4); } - len = t; } out_shift_with_cnt ("lsr %B0" CR_TAB - "ror %A0", insn, operands, len, 2); + "ror %A0", insn, operands, plen, 2); return ""; } @@ -7405,34 +7641,162 @@ lshrhi3_out (rtx_insn *insn, rtx operands[], int *len) /* 24-bit logic shift right */ const char* -avr_out_lshrpsi3 (rtx_insn *insn, rtx *op, int *plen) +avr_out_lshrpsi3 (rtx_insn *insn, rtx *operands, int *plen) { + rtx op[5]; + + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; + int dest = REGNO (op[0]); int src = REGNO (op[1]); if (CONST_INT_P (op[2])) { - if (plen) - *plen = 0; + int scratch = (GET_CODE (PATTERN (insn)) == PARALLEL + && XVECLEN (PATTERN (insn), 0) == 3 + && REG_P (operands[3])); + int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); switch (INTVAL (op[2])) { + + case 4: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + if (ldi_ok) + { + return avr_asm_len ("swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %A0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "andi %B0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "andi %C0,0x0f" CR_TAB + "eor %B0,%C0", op, plen, 10); + } + if (scratch) + { + return avr_asm_len ("swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0x0f" CR_TAB + "and %A0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "and %B0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "and %C0,%3" CR_TAB + "eor %B0,%C0", op, plen, 11); + } + break; + + case 5: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + if (ldi_ok) + { + return avr_asm_len ("lsr %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %A0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "andi %B0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "andi %C0,0x0f" CR_TAB + "eor %B0,%C0", op, plen, 13); + } + if (scratch) + { + return avr_asm_len ("lsr %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0x0f" CR_TAB + "and %A0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "and %B0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "and %C0,%3" CR_TAB + "eor %B0,%C0", op, plen, 14); + } + break; /* 10 */ + + case 6: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + return avr_asm_len ("clr __tmp_reg__" CR_TAB + "lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol __tmp_reg__" CR_TAB + "lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol __tmp_reg__" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "mov %C0,__tmp_reg__", op, plen, 12); + + case 7: + if (optimize_size) + break; /* scratch ? 6 : 7 */ + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "sbc %C0,%C0" CR_TAB + "neg %C0", op, plen, 7); + case 8: - if (dest <= src) - return avr_asm_len ("mov %A0,%B1" CR_TAB - "mov %B0,%C1" CR_TAB - "clr %C0", op, plen, 3); - else - return avr_asm_len ("clr %C0" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %A0,%B1", op, plen, 3); + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: + { + if (dest <= src) + avr_asm_len("mov %A0,%B1" CR_TAB + "mov %B0,%C1" CR_TAB + "clr %C0", op, plen, 3); + else + avr_asm_len("clr %C0" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %A0,%B1", op, plen, 3); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return lshrhi3_out(insn, op, plen); + } case 16: - if (dest != src + 2) - avr_asm_len ("mov %A0,%C1", op, plen, 1); + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: + { + if (dest != src + 2) + avr_asm_len("mov %A0,%C1", op, plen, 1); - return avr_asm_len ("clr %B0" CR_TAB - "clr %C0", op, plen, 2); + avr_asm_len("clr %B0" CR_TAB + "clr %C0", op, plen, 2); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return lshrqi3_out(insn, op, plen); + } default: if (INTVAL (op[2]) < 24) @@ -7459,15 +7823,21 @@ avr_out_lshrpsi3 (rtx_insn *insn, rtx *op, int *plen) /* 32-bit logic shift right ((unsigned int)x >> i) */ const char * -lshrsi3_out (rtx_insn *insn, rtx operands[], int *len) +lshrsi3_out (rtx_insn *insn, rtx operands[], int *plen) { + rtx op[5]; + + op[0] = operands[0]; + op[1] = operands[1]; + op[2] = operands[2]; + op[3] = operands[3]; + if (CONST_INT_P (operands[2])) { - int k; - int *t = len; - - if (!len) - len = &k; + int scratch = (GET_CODE (PATTERN (insn)) == PARALLEL + && XVECLEN (PATTERN (insn), 0) == 3 + && REG_P (operands[3])); + int ldi_ok = test_hard_reg_class (LD_REGS, operands[0]); switch (INTVAL (operands[2])) { @@ -7476,72 +7846,217 @@ lshrsi3_out (rtx_insn *insn, rtx operands[], int *len) break; if (AVR_HAVE_MOVW) - return *len = 3, ("clr %D0" CR_TAB + return avr_asm_len ("clr %D0" CR_TAB "clr %C0" CR_TAB - "movw %A0,%C0"); - *len = 4; - return ("clr %D0" CR_TAB + "movw %A0,%C0", operands, plen, 3); + + return avr_asm_len ("clr %D0" CR_TAB "clr %C0" CR_TAB "clr %B0" CR_TAB - "clr %A0"); + "clr %A0", operands, plen, 4); - case 8: - { + case 4: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + if (ldi_ok) + { + return avr_asm_len ("swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %A0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "andi %B0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "andi %C0,0x0f" CR_TAB + "eor %B0,%C0" CR_TAB + "eor %C0,%D0" CR_TAB + "andi %D0,0x0f" CR_TAB + "eor %C0,%D0", op, plen, 14); + } + if (scratch) + { + return avr_asm_len ("swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0x0f" CR_TAB + "and %A0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "and %B0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "and %C0,%3" CR_TAB + "eor %B0,%C0" CR_TAB + "eor %C0,%D0" CR_TAB + "and %D0,%3" CR_TAB + "eor %C0,%D0", op, plen, 15); + } + break; + + case 5: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + if (ldi_ok) + { + return avr_asm_len ("lsr %D0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "andi %A0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "andi %B0,0x0f" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "andi %C0,0x0f" CR_TAB + "eor %B0,%C0" CR_TAB + "eor %C0,%D0" CR_TAB + "andi %D0,0x0f" CR_TAB + "eor %C0,%D0", op, plen, 18); + } + if (scratch) + { + return avr_asm_len ("lsr %D0" CR_TAB + "ror %C0" CR_TAB + "ror %B0" CR_TAB + "ror %A0" CR_TAB + "swap %D0" CR_TAB + "swap %C0" CR_TAB + "swap %B0" CR_TAB + "swap %A0" CR_TAB + "ldi %3,0x0f" CR_TAB + "and %A0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "and %B0,%3" CR_TAB + "eor %A0,%B0" CR_TAB + "eor %B0,%C0" CR_TAB + "and %C0,%3" CR_TAB + "eor %B0,%C0" CR_TAB + "eor %C0,%D0" CR_TAB + "and %D0,%3" CR_TAB + "eor %C0,%D0", op, plen, 19); + } + break; /* 10 */ + + case 6: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + return avr_asm_len ("clr __tmp_reg__" CR_TAB + "lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "rol __tmp_reg__" CR_TAB + "lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "rol __tmp_reg__" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "mov %C0,%D0" CR_TAB + "mov %D0,__tmp_reg__", op, plen, 15); + + case 7: + if (optimize_size) + break; /* scratch ? 7 : 8 */ + return avr_asm_len ("lsl %A0" CR_TAB + "rol %B0" CR_TAB + "rol %C0" CR_TAB + "rol %D0" CR_TAB + "mov %A0,%B0" CR_TAB + "mov %B0,%C0" CR_TAB + "mov %C0,%D0" CR_TAB + "sbc %D0,%D0" CR_TAB + "neg %D0", op, plen, 9); + + case 8: + case 9: + case 10: + case 11: + case 12: + case 13: + case 14: + case 15: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); - *len = 4; if (reg0 <= reg1) - return ("mov %A0,%B1" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %C0,%D1" CR_TAB - "clr %D0"); + avr_asm_len("mov %A0,%B1" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %C0,%D1" CR_TAB + "clr %D0", op, plen, 4); else - return ("clr %D0" CR_TAB - "mov %C0,%D1" CR_TAB - "mov %B0,%C1" CR_TAB - "mov %A0,%B1"); - } - - case 16: - { + avr_asm_len("clr %D0" CR_TAB + "mov %C0,%D1" CR_TAB + "mov %B0,%C1" CR_TAB + "mov %A0,%B1", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 8); + return avr_out_lshrpsi3(insn, op, plen); + } + + case 16: + case 17: + case 18: + case 19: + case 20: + case 21: + case 22: + case 23: + { int reg0 = true_regnum (operands[0]); int reg1 = true_regnum (operands[1]); if (reg0 == reg1 + 2) - return *len = 2, ("clr %C0" CR_TAB - "clr %D0"); + avr_asm_len("clr %C0" CR_TAB + "clr %D0", op, plen, 2); if (AVR_HAVE_MOVW) - return *len = 3, ("movw %A0,%C1" CR_TAB - "clr %C0" CR_TAB - "clr %D0"); + avr_asm_len("movw %A0,%C1" CR_TAB + "clr %C0" CR_TAB + "clr %D0", op, plen, 3); else - return *len = 4, ("mov %B0,%D1" CR_TAB - "mov %A0,%C1" CR_TAB - "clr %C0" CR_TAB - "clr %D0"); - } - - case 24: - return *len = 4, ("mov %A0,%D1" CR_TAB - "clr %B0" CR_TAB - "clr %C0" CR_TAB - "clr %D0"); - - case 31: - *len = 6; - return ("clr %A0" CR_TAB + avr_asm_len("mov %B0,%D1" CR_TAB + "mov %A0,%C1" CR_TAB + "clr %C0" CR_TAB + "clr %D0", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 16); + return lshrhi3_out(insn, op, plen); + } + + case 24: + case 25: + case 26: + case 27: + case 28: + case 29: + case 30: + { + avr_asm_len("mov %A0,%D1" CR_TAB + "clr %B0" CR_TAB + "clr %C0" CR_TAB + "clr %D0", op, plen, 4); + op[2] = GEN_INT((INTVAL (operands[2])) - 24); + return lshrqi3_out(insn, op, plen); + } + + case 31: + return avr_asm_len ("clr %A0" CR_TAB "sbrc %D0,7" CR_TAB "inc %A0" CR_TAB "clr %B0" CR_TAB "clr %C0" CR_TAB - "clr %D0"); + "clr %D0", operands, plen, 6); } - len = t; } out_shift_with_cnt ("lsr %D0" CR_TAB "ror %C0" CR_TAB "ror %B0" CR_TAB - "ror %A0", insn, operands, len, 4); + "ror %A0", insn, operands, plen, 4); return ""; }
Comments
On 10/4/22 11:06, Alexander Binzberger via Gcc-patches wrote: > Hi, > recently I used some arduino uno for a project and realized some areas > which do not output optimal asm code. Especially around shifts and function > calls. > With this as motivation and hacktoberfest I started patching things. > Since patch files do not provide a good overview and I hope for a > "hacktoberfest-accepted" label on the PR on github I also opened it there: > https://github.com/gcc-mirror/gcc/pull/73 > > This patch improves shifts with const right hand operand. While 8bit and > 16bit shifts where mostly fine 24bit and 32bit where not handled well. > > Testing > I checked output with a local installation of compiler explorer in asm and > a tiny unit test comparing shifts with mul/div by 2. > I however did not write any testcases in gcc for it. > > Target > This patch is only targeting atmel avr family of chips. > > Changelog > improved const shifts for AVR targets It would be helpful if you could show the before/after code for the cases you're changing. Extra credit if you include cycles & size information for those cases. That would help someone like me who knows GCC well, but isn't particularly well versed in the AVR target evaluate the overarching goal of the patch (ie, better code). Changes should include a ChangeLog which indicates what changed. If you look at git log you will see examples of what a ChangeLog should look like. The is large enough that you need either a copyright assignment or DCO certification. See this page for details: https://gcc.gnu.org/contribute.html > > Patch > ----- > diff --git a/gcc/config/avr/avr.cc b/gcc/config/avr/avr.cc > index 4ed390e4cf9..c7b70812d5c 100644 > --- a/gcc/config/avr/avr.cc > +++ b/gcc/config/avr/avr.cc > @@ -6043,9 +6043,6 @@ out_shift_with_cnt (const char *templ, rtx_insn > *insn, rtx operands[], > op[2] = operands[2]; > op[3] = operands[3]; > > - if (plen) > - *plen = 0; > - Doesn't this leave *plen uninitialized for the case where the shift count is held in memory or a register or is an out of range constant? Is this really safe? > if (CONST_INT_P (operands[2])) > { > /* Operand 3 is a scratch register if this is a > @@ -6150,96 +6147,68 @@ out_shift_with_cnt (const char *templ, rtx_insn > *insn, rtx operands[], > /* 8bit shift left ((char)x << i) */ > > const char * > -ashlqi3_out (rtx_insn *insn, rtx operands[], int *len) > +ashlqi3_out (rtx_insn *insn, rtx operands[], int *plen) > { > if (CONST_INT_P (operands[2])) > { > - int k; > - > - if (!len) > - len = &k; > - Isn't this wrong for the call to ashlqi3_out from avr.md? In that call len/plen will be zero, which we then pass down. So the question is why did you remove this code? The patch as-is is relatively large and can easily be broken down into more manageable chunks. I would suggest a patch for each mode. ie, one which changes QImode shifts, another for HImode shifts, another for PSImode shifts. etc. It may seem like more work, but by breaking it down reviewers can take action on each patch individually. So for example its relatively easy to work through the QImode changes and those could go in fairly quick while the PSImode changes will require considerably more time to review. > switch (INTVAL (operands[2])) > { > default: > if (INTVAL (operands[2]) < 8) > break; > > - *len = 1; > - return "clr %0"; > - > - case 1: > - *len = 1; > - return "lsl %0"; > - > - case 2: > - *len = 2; > - return ("lsl %0" CR_TAB > - "lsl %0"); > - > - case 3: > - *len = 3; > - return ("lsl %0" CR_TAB > - "lsl %0" CR_TAB > - "lsl %0"); > + return avr_asm_len ("clr %0", operands, plen, 1); You've probably got a whitespace problem here. I think the return should line up in the came column as the IF statement. Conceptually this seems reasonable as cases 1, 2 and 3 can be trivially handled by out_shift_with_cnt. Tough routing more code through out_shift_with_cnt means the comment might need to change since we're routing more cases through it that were trivially handled in ashlqi3_out before. > > case 4: > if (test_hard_reg_class (LD_REGS, operands[0])) > { > - *len = 2; > - return ("swap %0" CR_TAB > - "andi %0,0xf0"); > + return avr_asm_len ("swap %0" CR_TAB > + "andi %0,0xf0", operands, plen, 2); More indention problems here. THe return should line up two spaces inside the open curly brace. Otherwise this case seems reasonable since it's generating the same code as before. > } > - *len = 4; > - return ("lsl %0" CR_TAB > + return avr_asm_len ("lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > - "lsl %0"); > + "lsl %0", operands, plen, 4); Gratuitous indentation changes. Please don't do that unless you're fixing cases where the indentation is wrong according to GNU/project standards. > > case 5: > if (test_hard_reg_class (LD_REGS, operands[0])) > { > - *len = 3; > - return ("swap %0" CR_TAB > + return avr_asm_len ("swap %0" CR_TAB > "lsl %0" CR_TAB > - "andi %0,0xe0"); > + "andi %0,0xe0", operands, plen, 3); It looks like you're introducing gratuitous indentation changes here. Please don't do that. Otherwise this looks sensible as again, it doesn't change the generated code. > } > - *len = 5; > - return ("lsl %0" CR_TAB > + return avr_asm_len ("lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > - "lsl %0"); > + "lsl %0", operands, plen, 5); Similarly. > > case 6: > if (test_hard_reg_class (LD_REGS, operands[0])) > { > - *len = 4; > - return ("swap %0" CR_TAB > + return avr_asm_len ("swap %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > - "andi %0,0xc0"); > + "andi %0,0xc0", operands, plen, 4); > } > - *len = 6; > - return ("lsl %0" CR_TAB > + return avr_asm_len ("lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > "lsl %0" CR_TAB > - "lsl %0"); > + "lsl %0", operands, plen, 6); > > case 7: > - *len = 3; > - return ("ror %0" CR_TAB > + return avr_asm_len ("ror %0" CR_TAB > "clr %0" CR_TAB > - "ror %0"); > + "ror %0", operands, plen, 3); > } Similarly for these cases. I don't have the time right now to go through the other modes. But I would suggest you take the feedback above and apply the same concepts to the changes for the other modes. Note for other reviewers, without an assignment or DCO cert, this can't go forward. jeff
On 12.10.22 19:57, Jeff Law wrote: > > On 10/4/22 11:06, Alexander Binzberger via Gcc-patches wrote: >> Hi, >> recently I used some arduino uno for a project and realized some areas >> which do not output optimal asm code. Especially around shifts and >> function >> calls. >> With this as motivation and hacktoberfest I started patching things. >> Since patch files do not provide a good overview and I hope for a >> "hacktoberfest-accepted" label on the PR on github I also opened it >> there: >> https://github.com/gcc-mirror/gcc/pull/73 >> >> This patch improves shifts with const right hand operand. While 8bit and >> 16bit shifts where mostly fine 24bit and 32bit where not handled well. >> >> Testing >> I checked output with a local installation of compiler explorer in >> asm and >> a tiny unit test comparing shifts with mul/div by 2. >> I however did not write any testcases in gcc for it. >> >> Target >> This patch is only targeting atmel avr family of chips. >> >> Changelog >> improved const shifts for AVR targets > > It would be helpful if you could show the before/after code for the > cases you're changing. Extra credit if you include cycles & size > information for those cases. That would help someone like me who > knows GCC well, but isn't particularly well versed in the AVR target > evaluate the overarching goal of the patch (ie, better code). about the avr family targets: * consider every branch instruction = 1/2 cycles * consider every 2byte/word instruction (besides move word if available) = 2 cycles * consider multiplication (if available) = 2 cycles * consider every load (beside load immediate "ldi" 1cylce) = 2cycles (+1 for prog mem) * pop and jump mostly = 2 cycles * call is basically = 2-4 cycles * ret is about = 4/5 cycles * consider every instruction (bit/bit-test, most compare, arithmetic, logic, some other) = 1 cycle * division does not exist or as a summary for this patch: branches and such are 2 cycles the rest is 1 cycle note that shifts are 1bit per cycle and the instructions are at least mostly byte based. also note that operations using immediate do only work with the upper half of registers. a description for the code before my change and what changed: * shifts on 8bit (beside arithmetic shifts right) were optimized and always unrolled (only aligned with the rest of the code without actual change) * arithmetic shift 8bit and 16bit shifts were mostly optimized and mostly unrolled - depending on registers and Os (I added the missing cases there) * 24bit and 32bit shifts were basically not optimized at all and never unrolled (I added those cases and aligned the optimizer logic with the others. They also reuse the other shift code since they may reduce to those cases after a move for bigger shifts.) * out_shift_with_cnt provides a fallback implementation as a loop over shifts which may get unrolled. in case of Os to about inner_len + 3,4 or 5 and in other cases of optimizer e.g. O2 it gets unrolled if size is smaller 10. see max_len (basically unchanged) * did not touch non const cases in this patch but may in a future patch for O2 and O3 note that in case of Os the smaller code is picked which is the loop at least in some cases but other optimizer cases profit a lot. also note that it is debatable if Os needs to be that strict with size since the compute overhead of the loop is high with 5 per loop iteration/cycle- so per bit shift. A lot more cases could be covered with +1 or +2 more instructions. about plen: If plen is NULL the asm code gets returned. If plen is a pointer the code does count the instruction count which I guess is used (or could be used) as a rough estimate of cycles as well as byte code size. Some of the functions named this len. The 24bit functions mainly named this plen and used it like it is now in all functions. This is mostly a readability improvement. I am not sure how this works together with the optimizer or the rest. To my understanding however the functions may get called once by the optimizer with a length given, then to output code and potentially again with a len given over avr_adjust_length to return the size. I may be wrong about this part but as far as I can tell I did not change the way it operates. size and cycles summary: The asm instruction count is used as size and cycle estimate. This gets close to the real deal for the shifts since the instructions are all 1 cylce anyway and similar in byte code size. 8bit gets always optimized and unrolled to get max performance and less code size (beside shift of 6 with lower half registers used which is the worst case with +1 instruction). 16bit, 24bit and 32bit gets unrolled depending on optimizer setting - and registers used (see out_shift_with_cnt:max_len). So 16bit gets mostly optimized and unrolled in Os (see comments for plen/max_len) and always in O2 and such (max_len=10). Shift optimization and unroll for 24bit and 32bit is mostly only relevant when not optimizing for size. I think the move (for shifts bigger than a multiple of 8 optimization) should always be done if possible since this is a tiny space overhead but a big performance gain. Also bigger shifts on bigger variables may not be common any way. > > Changes should include a ChangeLog which indicates what changed. If > you look at git log you will see examples of what a ChangeLog should > look like. are you fine with something like the following? one line for all functions then, same text? gcc/ChangeLog: * config/avr/avr.cc (ashlqi3_out): improved const shifts optimization > > The is large enough that you need either a copyright assignment or > DCO certification. > > See this page for details: > > https://gcc.gnu.org/contribute.html Signed-off-by: Alexander Binzberger <alexander.binzberger@gmail.com> <mailto:alexander.binzberger@gmail.com> Is this enough or did I miss something? > > >> >> Patch >> ----- >> diff --git a/gcc/config/avr/avr.cc b/gcc/config/avr/avr.cc >> index 4ed390e4cf9..c7b70812d5c 100644 >> --- a/gcc/config/avr/avr.cc >> +++ b/gcc/config/avr/avr.cc >> @@ -6043,9 +6043,6 @@ out_shift_with_cnt (const char *templ, rtx_insn >> *insn, rtx operands[], >> op[2] = operands[2]; >> op[3] = operands[3]; >> >> - if (plen) >> - *plen = 0; >> - > > Doesn't this leave *plen uninitialized for the case where the shift > count is held in memory or a register or is an out of range constant? > Is this really safe? To my understanding this is basically how the code was working before - especially in the 24bit functions. It checks for NULL in avr_asm_len function which is why I use that now everywhere. This is why this looks fine to me but more eyes on this is a good idea ;) Since I am not sure how this plen gets used in other parts of the code or by the optimizer I wonder if it should get set to 0 at all at this place. I guess this must happen somewhere up in the call tree anyway and should maybe not get set there. This however is consistent with the rest of the code in this file as far as I can tell. I would be very happy if you could have a second look at this topic. > > > >> if (CONST_INT_P (operands[2])) >> { >> /* Operand 3 is a scratch register if this is a >> @@ -6150,96 +6147,68 @@ out_shift_with_cnt (const char *templ, rtx_insn >> *insn, rtx operands[], >> /* 8bit shift left ((char)x << i) */ >> >> const char * >> -ashlqi3_out (rtx_insn *insn, rtx operands[], int *len) >> +ashlqi3_out (rtx_insn *insn, rtx operands[], int *plen) >> { >> if (CONST_INT_P (operands[2])) >> { >> - int k; >> - >> - if (!len) >> - len = &k; >> - > > Isn't this wrong for the call to ashlqi3_out from avr.md? In that > call len/plen will be zero, which we then pass down. So the question > is why did you remove this code? About half of the code was written with plen and avr_asm_len the other parts like this. The plen variation was the more flexible one and made some changes down the road more easy so I made it use this one consistently. > > > The patch as-is is relatively large and can easily be broken down into > more manageable chunks. I would suggest a patch for each mode. ie, > one which changes QImode shifts, another for HImode shifts, another > for PSImode shifts. etc. It may seem like more work, but by breaking > it down reviewers can take action on each patch individually. So for > example its relatively easy to work through the QImode changes and > those could go in fairly quick while the PSImode changes will require > considerably more time to review. Due to the plen change (needed for some readability improvements, more consistent code but also more simple changes) and other minor readability improvements I decided to go with one patch for all const cases. Are you sure you want this extra work and overhead? Would you prefer them as a set of patches/mails (with patch 1/x) or single patch mails? Maybe as github pull requests with multiple commits? I am motivated to do what fits you best. I will fix the spacing after your answer to this. > > >> switch (INTVAL (operands[2])) >> { >> default: >> if (INTVAL (operands[2]) < 8) >> break; >> >> - *len = 1; >> - return "clr %0"; >> - >> - case 1: >> - *len = 1; >> - return "lsl %0"; >> - >> - case 2: >> - *len = 2; >> - return ("lsl %0" CR_TAB >> - "lsl %0"); >> - >> - case 3: >> - *len = 3; >> - return ("lsl %0" CR_TAB >> - "lsl %0" CR_TAB >> - "lsl %0"); >> + return avr_asm_len ("clr %0", operands, plen, 1); > > You've probably got a whitespace problem here. I think the return > should line up in the came column as the IF statement. yes, this tab=8 and is mixed with spaces thing confused me and the editor I think. Will fix this with the next iteration, see above comment. > Conceptually this seems reasonable as cases 1, 2 and 3 can be > trivially handled by out_shift_with_cnt. Tough routing more code > through out_shift_with_cnt means the comment might need to change > since we're routing more cases through it that were trivially handled > in ashlqi3_out before. This is one of the readability improvements. There is no optimization for cases 1-3 I know of and out_shift_with_cnt will unroll them anyway in those cases resulting in the same asm code. Maybe I should add a comment there. > > >> >> case 4: >> if (test_hard_reg_class (LD_REGS, operands[0])) >> { >> - *len = 2; >> - return ("swap %0" CR_TAB >> - "andi %0,0xf0"); >> + return avr_asm_len ("swap %0" CR_TAB >> + "andi %0,0xf0", operands, plen, 2); > More indention problems here. THe return should line up two spaces > inside the open curly brace. Otherwise this case seems reasonable > since it's generating the same code as before. see plen change explanation. >> } >> - *len = 4; >> - return ("lsl %0" CR_TAB >> + return avr_asm_len ("lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> - "lsl %0"); >> + "lsl %0", operands, plen, 4); > > Gratuitous indentation changes. Please don't do that unless you're > fixing cases where the indentation is wrong according to GNU/project > standards. Will fix this with next iteration, see above comment. > > >> >> case 5: >> if (test_hard_reg_class (LD_REGS, operands[0])) >> { >> - *len = 3; >> - return ("swap %0" CR_TAB >> + return avr_asm_len ("swap %0" CR_TAB >> "lsl %0" CR_TAB >> - "andi %0,0xe0"); >> + "andi %0,0xe0", operands, plen, 3); > > It looks like you're introducing gratuitous indentation changes here. > Please don't do that. Otherwise this looks sensible as again, it > doesn't change the generated code. > > > >> } >> - *len = 5; >> - return ("lsl %0" CR_TAB >> + return avr_asm_len ("lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> - "lsl %0"); >> + "lsl %0", operands, plen, 5); > > Similarly. > > >> >> case 6: >> if (test_hard_reg_class (LD_REGS, operands[0])) >> { >> - *len = 4; >> - return ("swap %0" CR_TAB >> + return avr_asm_len ("swap %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> - "andi %0,0xc0"); >> + "andi %0,0xc0", operands, plen, 4); >> } >> - *len = 6; >> - return ("lsl %0" CR_TAB >> + return avr_asm_len ("lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> "lsl %0" CR_TAB >> - "lsl %0"); >> + "lsl %0", operands, plen, 6); >> >> case 7: >> - *len = 3; >> - return ("ror %0" CR_TAB >> + return avr_asm_len ("ror %0" CR_TAB >> "clr %0" CR_TAB >> - "ror %0"); >> + "ror %0", operands, plen, 3); >> } > > Similarly for these cases. > > > I don't have the time right now to go through the other modes. But I > would suggest you take the feedback above and apply the same concepts > to the changes for the other modes. Thanks for your time! > > > Note for other reviewers, without an assignment or DCO cert, this > can't go forward. see sign-off above. is this fine? something else needed? > > > jeff >
On 10/15/22 06:08, A. Binzberger wrote: > Re: [PATCH] improved const shifts for AVR targets > On 12.10.22 19:57, Jeff Law wrote: >> >> On 10/4/22 11:06, Alexander Binzberger via Gcc-patches wrote: >>> Hi, >>> recently I used some arduino uno for a project and realized some areas >>> which do not output optimal asm code. Especially around shifts and >>> function >>> calls. >>> With this as motivation and hacktoberfest I started patching things. >>> Since patch files do not provide a good overview and I hope for a >>> "hacktoberfest-accepted" label on the PR on github I also opened it >>> there: >>> https://github.com/gcc-mirror/gcc/pull/73 >>> >>> This patch improves shifts with const right hand operand. While 8bit >>> and >>> 16bit shifts where mostly fine 24bit and 32bit where not handled well. >>> >>> Testing >>> I checked output with a local installation of compiler explorer in >>> asm and >>> a tiny unit test comparing shifts with mul/div by 2. >>> I however did not write any testcases in gcc for it. >>> >>> Target >>> This patch is only targeting atmel avr family of chips. >>> >>> Changelog >>> improved const shifts for AVR targets >> >> It would be helpful if you could show the before/after code for the >> cases you're changing. Extra credit if you include cycles & size >> information for those cases. That would help someone like me who >> knows GCC well, but isn't particularly well versed in the AVR target >> evaluate the overarching goal of the patch (ie, better code). > > about the avr family targets: > > * consider every branch instruction = 1/2 cycles > > * consider every 2byte/word instruction (besides move word if > available) = 2 cycles > > * consider multiplication (if available) = 2 cycles > > * consider every load (beside load immediate "ldi" 1cylce) = 2cycles > (+1 for prog mem) > > * pop and jump mostly = 2 cycles > > * call is basically = 2-4 cycles > > * ret is about = 4/5 cycles > > * consider every instruction (bit/bit-test, most compare, arithmetic, > logic, some other) = 1 cycle > > * division does not exist > > or as a summary for this patch: branches and such are 2 cycles the > rest is 1 cycle > > note that shifts are 1bit per cycle and the instructions are at least > mostly byte based. > > also note that operations using immediate do only work with the upper > half of registers. > All useful, but you should be giving me the summary for the things you're changing, not asking me to do it :-) Presumably you've already done the analysis to ensure your changes are an improvement. I'm asking you to provide that analysis for review and archival purposes. A quick table like Mode Shift count Shift type original cycles (or size) new cycles (or size) That will make it very clear for me and anyone doing historical work in the future what was expected here. It's OK if the cycle counts aren't 100% accurate. Including a testcase would be awesome as well, but isn't strictly required. > > a description for the code before my change and what changed: > > * shifts on 8bit (beside arithmetic shifts right) were optimized and > always unrolled (only aligned with the rest of the code without actual > change) > > * arithmetic shift 8bit and 16bit shifts were mostly optimized and > mostly unrolled - depending on registers and Os (I added the missing > cases there) > > * 24bit and 32bit shifts were basically not optimized at all and never > unrolled (I added those cases and aligned the optimizer logic with the > others. They also reuse the other shift code since they may reduce to > those cases after a move for bigger shifts.) > > * out_shift_with_cnt provides a fallback implementation as a loop over > shifts which may get unrolled. in case of Os to about inner_len + 3,4 > or 5 and in other cases of optimizer e.g. O2 it gets unrolled if size > is smaller 10. see max_len (basically unchanged) > > * did not touch non const cases in this patch but may in a future > patch for O2 and O3 > > note that in case of Os the smaller code is picked which is the loop > at least in some cases but other optimizer cases profit a lot. > > also note that it is debatable if Os needs to be that strict with size > since the compute overhead of the loop is high with 5 per loop > iteration/cycle- so per bit shift. A lot more cases could be covered > with +1 or +2 more instructions. > > > about plen: > > If plen is NULL the asm code gets returned. > > If plen is a pointer the code does count the instruction count which I > guess is used (or could be used) as a rough estimate of cycles as well > as byte code size. > > Some of the functions named this len. The 24bit functions mainly named > this plen and used it like it is now in all functions. This is mostly > a readability improvement. > > I am not sure how this works together with the optimizer or the rest. > > To my understanding however the functions may get called once by the > optimizer with a length given, then to output code and potentially > again with a len given over avr_adjust_length to return the size. > > I may be wrong about this part but as far as I can tell I did not > change the way it operates. > > > size and cycles summary: > > The asm instruction count is used as size and cycle estimate. This > gets close to the real deal for the shifts since the instructions are > all 1 cylce anyway and similar in byte code size. > > 8bit gets always optimized and unrolled to get max performance and > less code size (beside shift of 6 with lower half registers used which > is the worst case with +1 instruction). > > 16bit, 24bit and 32bit gets unrolled depending on optimizer setting - > and registers used (see out_shift_with_cnt:max_len). So 16bit gets > mostly optimized and unrolled in Os (see comments for plen/max_len) > and always in O2 and such (max_len=10). Shift optimization and unroll > for 24bit and 32bit is mostly only relevant when not optimizing for size. > > I think the move (for shifts bigger than a multiple of 8 optimization) > should always be done if possible since this is a tiny space overhead > but a big performance gain. Also bigger shifts on bigger variables may > not be common any way. > >> >> Changes should include a ChangeLog which indicates what changed. If >> you look at git log you will see examples of what a ChangeLog should >> look like. > > are you fine with something like the following? one line for all > functions then, same text? > > gcc/ChangeLog: > > * config/avr/avr.cc (ashlqi3_out): improved const shifts optimization > A ChangeLog should have an entry for each function that was changed and a brief description of what changed. Generally start them with a capitol and end with a period. So for that one * config/avr/avr.cc (ashlqi3_out): Improve shift by constants. But you need an entry for each function you changed. >> >> The is large enough that you need either a copyright assignment or >> DCO certification. >> >> See this page for details: >> >> https://gcc.gnu.org/contribute.html > > Signed-off-by: Alexander Binzberger <alexander.binzberger@gmail.com> > <mailto:alexander.binzberger@gmail.com> > > Is this enough or did I miss something? > That's sufficient. Thanks. >> >> >>> >>> Patch >>> ----- >>> diff --git a/gcc/config/avr/avr.cc b/gcc/config/avr/avr.cc >>> index 4ed390e4cf9..c7b70812d5c 100644 >>> --- a/gcc/config/avr/avr.cc >>> +++ b/gcc/config/avr/avr.cc >>> @@ -6043,9 +6043,6 @@ out_shift_with_cnt (const char *templ, rtx_insn >>> *insn, rtx operands[], >>> op[2] = operands[2]; >>> op[3] = operands[3]; >>> >>> - if (plen) >>> - *plen = 0; >>> - >> >> Doesn't this leave *plen uninitialized for the case where the shift >> count is held in memory or a register or is an out of range >> constant? Is this really safe? > > To my understanding this is basically how the code was working before > - especially in the 24bit functions. It checks for NULL in avr_asm_len > function which is why I use that now everywhere. This is why this > looks fine to me but more eyes on this is a good idea ;) > > Since I am not sure how this plen gets used in other parts of the code > or by the optimizer I wonder if it should get set to 0 at all at this > place. I guess this must happen somewhere up in the call tree anyway > and should maybe not get set there. This however is consistent with > the rest of the code in this file as far as I can tell. > > I would be very happy if you could have a second look at this topic. > I think Georg commented that this change was incorrect as well. In particular his comment indicated you will get garbage for the instruction lengths, which in turn can cause runtime or linker errors. Georg's comments should be considered authoritative for changes to the AVR port. >> >> >> >>> if (CONST_INT_P (operands[2])) >>> { >>> /* Operand 3 is a scratch register if this is a >>> @@ -6150,96 +6147,68 @@ out_shift_with_cnt (const char *templ, rtx_insn >>> *insn, rtx operands[], >>> /* 8bit shift left ((char)x << i) */ >>> >>> const char * >>> -ashlqi3_out (rtx_insn *insn, rtx operands[], int *len) >>> +ashlqi3_out (rtx_insn *insn, rtx operands[], int *plen) >>> { >>> if (CONST_INT_P (operands[2])) >>> { >>> - int k; >>> - >>> - if (!len) >>> - len = &k; >>> - >> >> Isn't this wrong for the call to ashlqi3_out from avr.md? In that >> call len/plen will be zero, which we then pass down. So the question >> is why did you remove this code? > About half of the code was written with plen and avr_asm_len the other > parts like this. The plen variation was the more flexible one and made > some changes down the road more easy so I made it use this one > consistently. But as Georg has pointed it, you've changed the behavior of the code, likely creating problems as a result. >> >> >> The patch as-is is relatively large and can easily be broken down >> into more manageable chunks. I would suggest a patch for each mode. >> ie, one which changes QImode shifts, another for HImode shifts, >> another for PSImode shifts. etc. It may seem like more work, but by >> breaking it down reviewers can take action on each patch >> individually. So for example its relatively easy to work through the >> QImode changes and those could go in fairly quick while the PSImode >> changes will require considerably more time to review. > > Due to the plen change (needed for some readability improvements, more > consistent code but also more simple changes) and other minor > readability improvements I decided to go with one patch for all const > cases. > > Are you sure you want this extra work and overhead? > Yes. It's standard procedure. It also means that if there are parts of the work that are non-controversial, then they can go in immediately while other parts are refined. The SImode and PSImode changes in particular will require a lot of time to evaluate for correctness. > > Would you prefer them as a set of patches/mails (with patch 1/x) or > single patch mails? > Most folks use [1/x] as a series of patches. > Maybe as github pull requests with multiple commits? > One day I'd love to be able to accept pull requests rather than having to do everything via email :-) BUt we're not really set up for that yet. Jeff
diff --git a/gcc/config/avr/avr.cc b/gcc/config/avr/avr.cc index 4ed390e4cf9..c7b70812d5c 100644 --- a/gcc/config/avr/avr.cc +++ b/gcc/config/avr/avr.cc @@ -6043,9 +6043,6 @@ out_shift_with_cnt (const char *templ, rtx_insn *insn, rtx operands[], op[2] = operands[2]; op[3] = operands[3]; - if (plen) - *plen = 0; - if (CONST_INT_P (operands[2])) {