Phil Edworthy | 8e5ab4b | 2011-06-02 22:15:27 +0000 | [diff] [blame] | 1 | /* Copyright (C) 1994, 1995, 1997, 1998, 1999, 2000, 2001, 2002, 2003, |
| 2 | 2004, 2005, 2006 |
| 3 | Free Software Foundation, Inc. |
| 4 | |
| 5 | This file is free software; you can redistribute it and/or modify it |
| 6 | under the terms of the GNU General Public License as published by the |
| 7 | Free Software Foundation; either version 2, or (at your option) any |
| 8 | later version. |
| 9 | |
| 10 | In addition to the permissions in the GNU General Public License, the |
| 11 | Free Software Foundation gives you unlimited permission to link the |
| 12 | compiled version of this file into combinations with other programs, |
| 13 | and to distribute those combinations without any restriction coming |
| 14 | from the use of this file. (The General Public License restrictions |
| 15 | do apply in other respects; for example, they cover modification of |
| 16 | the file, and distribution when not linked into a combine |
| 17 | executable.) |
| 18 | |
| 19 | This file is distributed in the hope that it will be useful, but |
| 20 | WITHOUT ANY WARRANTY; without even the implied warranty of |
| 21 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 22 | General Public License for more details. |
| 23 | |
| 24 | You should have received a copy of the GNU General Public License |
| 25 | along with this program; see the file COPYING. If not, write to |
| 26 | the Free Software Foundation, 51 Franklin Street, Fifth Floor, |
| 27 | Boston, MA 02110-1301, USA. */ |
| 28 | |
| 29 | !! libgcc routines for the Renesas / SuperH SH CPUs. |
| 30 | !! Contributed by Steve Chamberlain. |
| 31 | !! sac@cygnus.com |
| 32 | |
| 33 | !! ashiftrt_r4_x, ___ashrsi3, ___ashlsi3, ___lshrsi3 routines |
| 34 | !! recoded in assembly by Toshiyasu Morita |
| 35 | !! tm@netcom.com |
| 36 | |
| 37 | /* SH2 optimizations for ___ashrsi3, ___ashlsi3, ___lshrsi3 and |
| 38 | ELF local label prefixes by J"orn Rennecke |
| 39 | amylaar@cygnus.com */ |
| 40 | |
| 41 | .text |
| 42 | .balign 4 |
| 43 | .global __movmem |
| 44 | .global __movstr |
| 45 | .set __movstr, __movmem |
| 46 | /* This would be a lot simpler if r6 contained the byte count |
| 47 | minus 64, and we wouldn't be called here for a byte count of 64. */ |
| 48 | __movmem: |
| 49 | sts.l pr,@-r15 |
| 50 | shll2 r6 |
| 51 | bsr __movmemSI52+2 |
| 52 | mov.l @(48,r5),r0 |
| 53 | .balign 4 |
| 54 | movmem_loop: /* Reached with rts */ |
| 55 | mov.l @(60,r5),r0 |
| 56 | add #-64,r6 |
| 57 | mov.l r0,@(60,r4) |
| 58 | tst r6,r6 |
| 59 | mov.l @(56,r5),r0 |
| 60 | bt movmem_done |
| 61 | mov.l r0,@(56,r4) |
| 62 | cmp/pl r6 |
| 63 | mov.l @(52,r5),r0 |
| 64 | add #64,r5 |
| 65 | mov.l r0,@(52,r4) |
| 66 | add #64,r4 |
| 67 | bt __movmemSI52 |
| 68 | ! done all the large groups, do the remainder |
| 69 | ! jump to movmem+ |
| 70 | mova __movmemSI4+4,r0 |
| 71 | add r6,r0 |
| 72 | jmp @r0 |
| 73 | movmem_done: ! share slot insn, works out aligned. |
| 74 | lds.l @r15+,pr |
| 75 | mov.l r0,@(56,r4) |
| 76 | mov.l @(52,r5),r0 |
| 77 | rts |
| 78 | mov.l r0,@(52,r4) |
| 79 | .balign 4 |
| 80 | |
| 81 | .global __movmemSI64 |
| 82 | .global __movstrSI64 |
| 83 | .set __movstrSI64, __movmemSI64 |
| 84 | __movmemSI64: |
| 85 | mov.l @(60,r5),r0 |
| 86 | mov.l r0,@(60,r4) |
| 87 | .global __movmemSI60 |
| 88 | .global __movstrSI60 |
| 89 | .set __movstrSI60, __movmemSI60 |
| 90 | __movmemSI60: |
| 91 | mov.l @(56,r5),r0 |
| 92 | mov.l r0,@(56,r4) |
| 93 | .global __movmemSI56 |
| 94 | .global __movstrSI56 |
| 95 | .set __movstrSI56, __movmemSI56 |
| 96 | __movmemSI56: |
| 97 | mov.l @(52,r5),r0 |
| 98 | mov.l r0,@(52,r4) |
| 99 | .global __movmemSI52 |
| 100 | .global __movstrSI52 |
| 101 | .set __movstrSI52, __movmemSI52 |
| 102 | __movmemSI52: |
| 103 | mov.l @(48,r5),r0 |
| 104 | mov.l r0,@(48,r4) |
| 105 | .global __movmemSI48 |
| 106 | .global __movstrSI48 |
| 107 | .set __movstrSI48, __movmemSI48 |
| 108 | __movmemSI48: |
| 109 | mov.l @(44,r5),r0 |
| 110 | mov.l r0,@(44,r4) |
| 111 | .global __movmemSI44 |
| 112 | .global __movstrSI44 |
| 113 | .set __movstrSI44, __movmemSI44 |
| 114 | __movmemSI44: |
| 115 | mov.l @(40,r5),r0 |
| 116 | mov.l r0,@(40,r4) |
| 117 | .global __movmemSI40 |
| 118 | .global __movstrSI40 |
| 119 | .set __movstrSI40, __movmemSI40 |
| 120 | __movmemSI40: |
| 121 | mov.l @(36,r5),r0 |
| 122 | mov.l r0,@(36,r4) |
| 123 | .global __movmemSI36 |
| 124 | .global __movstrSI36 |
| 125 | .set __movstrSI36, __movmemSI36 |
| 126 | __movmemSI36: |
| 127 | mov.l @(32,r5),r0 |
| 128 | mov.l r0,@(32,r4) |
| 129 | .global __movmemSI32 |
| 130 | .global __movstrSI32 |
| 131 | .set __movstrSI32, __movmemSI32 |
| 132 | __movmemSI32: |
| 133 | mov.l @(28,r5),r0 |
| 134 | mov.l r0,@(28,r4) |
| 135 | .global __movmemSI28 |
| 136 | .global __movstrSI28 |
| 137 | .set __movstrSI28, __movmemSI28 |
| 138 | __movmemSI28: |
| 139 | mov.l @(24,r5),r0 |
| 140 | mov.l r0,@(24,r4) |
| 141 | .global __movmemSI24 |
| 142 | .global __movstrSI24 |
| 143 | .set __movstrSI24, __movmemSI24 |
| 144 | __movmemSI24: |
| 145 | mov.l @(20,r5),r0 |
| 146 | mov.l r0,@(20,r4) |
| 147 | .global __movmemSI20 |
| 148 | .global __movstrSI20 |
| 149 | .set __movstrSI20, __movmemSI20 |
| 150 | __movmemSI20: |
| 151 | mov.l @(16,r5),r0 |
| 152 | mov.l r0,@(16,r4) |
| 153 | .global __movmemSI16 |
| 154 | .global __movstrSI16 |
| 155 | .set __movstrSI16, __movmemSI16 |
| 156 | __movmemSI16: |
| 157 | mov.l @(12,r5),r0 |
| 158 | mov.l r0,@(12,r4) |
| 159 | .global __movmemSI12 |
| 160 | .global __movstrSI12 |
| 161 | .set __movstrSI12, __movmemSI12 |
| 162 | __movmemSI12: |
| 163 | mov.l @(8,r5),r0 |
| 164 | mov.l r0,@(8,r4) |
| 165 | .global __movmemSI8 |
| 166 | .global __movstrSI8 |
| 167 | .set __movstrSI8, __movmemSI8 |
| 168 | __movmemSI8: |
| 169 | mov.l @(4,r5),r0 |
| 170 | mov.l r0,@(4,r4) |
| 171 | .global __movmemSI4 |
| 172 | .global __movstrSI4 |
| 173 | .set __movstrSI4, __movmemSI4 |
| 174 | __movmemSI4: |
| 175 | mov.l @(0,r5),r0 |
| 176 | rts |
| 177 | mov.l r0,@(0,r4) |
| 178 | |
| 179 | .global __movmem_i4_even |
| 180 | .global __movstr_i4_even |
| 181 | .set __movstr_i4_even, __movmem_i4_even |
| 182 | |
| 183 | .global __movmem_i4_odd |
| 184 | .global __movstr_i4_odd |
| 185 | .set __movstr_i4_odd, __movmem_i4_odd |
| 186 | |
| 187 | .global __movmemSI12_i4 |
| 188 | .global __movstrSI12_i4 |
| 189 | .set __movstrSI12_i4, __movmemSI12_i4 |
| 190 | |
| 191 | .p2align 5 |
| 192 | L_movmem_2mod4_end: |
| 193 | mov.l r0,@(16,r4) |
| 194 | rts |
| 195 | mov.l r1,@(20,r4) |
| 196 | |
| 197 | .p2align 2 |
| 198 | |
| 199 | __movmem_i4_even: |
| 200 | mov.l @r5+,r0 |
| 201 | bra L_movmem_start_even |
| 202 | mov.l @r5+,r1 |
| 203 | |
| 204 | __movmem_i4_odd: |
| 205 | mov.l @r5+,r1 |
| 206 | add #-4,r4 |
| 207 | mov.l @r5+,r2 |
| 208 | mov.l @r5+,r3 |
| 209 | mov.l r1,@(4,r4) |
| 210 | mov.l r2,@(8,r4) |
| 211 | |
| 212 | L_movmem_loop: |
| 213 | mov.l r3,@(12,r4) |
| 214 | dt r6 |
| 215 | mov.l @r5+,r0 |
| 216 | bt/s L_movmem_2mod4_end |
| 217 | mov.l @r5+,r1 |
| 218 | add #16,r4 |
| 219 | L_movmem_start_even: |
| 220 | mov.l @r5+,r2 |
| 221 | mov.l @r5+,r3 |
| 222 | mov.l r0,@r4 |
| 223 | dt r6 |
| 224 | mov.l r1,@(4,r4) |
| 225 | bf/s L_movmem_loop |
| 226 | mov.l r2,@(8,r4) |
| 227 | rts |
| 228 | mov.l r3,@(12,r4) |
| 229 | |
| 230 | .p2align 4 |
| 231 | __movmemSI12_i4: |
| 232 | mov.l @r5,r0 |
| 233 | mov.l @(4,r5),r1 |
| 234 | mov.l @(8,r5),r2 |
| 235 | mov.l r0,@r4 |
| 236 | mov.l r1,@(4,r4) |
| 237 | rts |
| 238 | mov.l r2,@(8,r4) |