1 /* Copyright (C) 2006-2014 Free Software Foundation, Inc.
2 This file is part of the GNU C Library.
4 Contributed by MontaVista Software, Inc. (written by Nicolas Pitre)
6 The GNU C Library is free software; you can redistribute it and/or
7 modify it under the terms of the GNU Lesser General Public
8 License as published by the Free Software Foundation; either
9 version 2.1 of the License, or (at your option) any later version.
11 The GNU C Library is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 Lesser General Public License for more details.
16 You should have received a copy of the GNU Lesser General Public
17 License along with the GNU C Library. If not, see
18 <http://www.gnu.org/licenses/>. */
20 /* Thumb requires excessive IT insns here. */
23 #include <arm-features.h>
26 * Data preload for architectures that support it (ARM V5TE and above)
28 #if (!defined (__ARM_ARCH_2__) && !defined (__ARM_ARCH_3__) \
29 && !defined (__ARM_ARCH_3M__) && !defined (__ARM_ARCH_4__) \
30 && !defined (__ARM_ARCH_4T__) && !defined (__ARM_ARCH_5__) \
31 && !defined (__ARM_ARCH_5T__))
32 #define PLD(code...) code
38 * This can be used to enable code to cacheline align the source pointer.
39 * Experiments on tested architectures (StrongARM and XScale) didn't show
40 * this a worthwhile thing to do. That might be different in the future.
42 //#define CALGN(code...) code
43 #define CALGN(code...)
46 * Endian independent macros for shifting bytes within registers.
60 * Prototype: void *memmove(void *dest, const void *src, size_t n);
64 * If the memory regions don't overlap, we simply branch to memcpy which is
65 * normally a bit faster. Otherwise the copy is done going downwards.
75 bls HIDDEN_JUMPTARGET(memcpy)
79 cfi_adjust_cfa_offset (12)
80 cfi_rel_offset (r4, 4)
81 cfi_rel_offset (lr, 8)
90 PLD( sfi_pld r1, #-4 )
97 cfi_adjust_cfa_offset (16)
98 cfi_rel_offset (r5, 0)
99 cfi_rel_offset (r6, 4)
100 cfi_rel_offset (r7, 8)
101 cfi_rel_offset (r8, 12)
104 CALGN( ands ip, r1, #31 )
105 CALGN( sbcsne r4, ip, r2 ) @ C is always set here
108 CALGN( subs r2, r2, ip ) @ C is set here
109 #ifndef ARM_ALWAYS_BX
110 CALGN( add pc, r4, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2))
112 CALGN( add r4, r4, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2))
116 PLD( sfi_pld r1, #-4 )
117 2: PLD( subs r2, r2, #96 )
118 PLD( sfi_pld r1, #-32 )
120 PLD( sfi_pld r1, #-64 )
121 PLD( sfi_pld r1, #-96 )
123 3: PLD( sfi_pld r1, #-128 )
125 ldmdb \B!, {r3, r4, r5, r6, r7, r8, ip, lr}
128 stmdb \B!, {r3, r4, r5, r6, r7, r8, ip, lr}
135 #ifndef ARM_ALWAYS_BX
136 /* C is always clear here. */
137 addne pc, pc, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2)
142 cfi_adjust_cfa_offset (4)
143 cfi_rel_offset (r10, 0)
144 add r10, pc, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2)
147 .p2align ARM_BX_ALIGN_LOG2
149 .p2align ARM_BX_ALIGN_LOG2
152 .p2align ARM_BX_ALIGN_LOG2
155 .p2align ARM_BX_ALIGN_LOG2
158 .p2align ARM_BX_ALIGN_LOG2
161 .p2align ARM_BX_ALIGN_LOG2
164 .p2align ARM_BX_ALIGN_LOG2
167 .p2align ARM_BX_ALIGN_LOG2
171 #ifndef ARM_ALWAYS_BX
172 add pc, pc, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2)
175 add r10, pc, ip, lsl #(ARM_BX_ALIGN_LOG2 - 2)
178 .p2align ARM_BX_ALIGN_LOG2
180 .p2align ARM_BX_ALIGN_LOG2
183 .p2align ARM_BX_ALIGN_LOG2
186 .p2align ARM_BX_ALIGN_LOG2
189 .p2align ARM_BX_ALIGN_LOG2
192 .p2align ARM_BX_ALIGN_LOG2
195 .p2align ARM_BX_ALIGN_LOG2
198 .p2align ARM_BX_ALIGN_LOG2
204 cfi_adjust_cfa_offset (-4)
211 cfi_adjust_cfa_offset (-16)
217 8: movs r2, r2, lsl #31
219 ldrbne r3, [\B, #-1]!
221 ldrbcs r4, [\B, #-1]!
225 strbne r3, [\B, #-1]!
227 strbcs r4, [\B, #-1]!
231 #if ((defined (__ARM_ARCH_4T__) && defined (__THUMB_INTERWORK__)) \
232 || defined (ARM_ALWAYS_BX))
234 cfi_adjust_cfa_offset (-12)
246 ldrbgt r3, [\B, #-1]!
248 ldrbge r4, [\B, #-1]!
252 strbgt r3, [\B, #-1]!
254 strbge r4, [\B, #-1]!
270 .macro backward_copy_shift push pull
275 CALGN( ands ip, r1, #31 )
276 CALGN( rsb ip, ip, #32 )
277 CALGN( sbcsne r4, ip, r2 ) @ C is always set here
278 CALGN( subcc r2, r2, ip )
281 11: push {r5 - r8, r10}
282 cfi_adjust_cfa_offset (20)
283 cfi_rel_offset (r5, 0)
284 cfi_rel_offset (r6, 4)
285 cfi_rel_offset (r7, 8)
286 cfi_rel_offset (r8, 12)
287 cfi_rel_offset (r10, 16)
289 PLD( sfi_pld r1, #-4 )
290 PLD( subs r2, r2, #96 )
291 PLD( sfi_pld r1, #-32 )
293 PLD( sfi_pld r1, #-64 )
294 PLD( sfi_pld r1, #-96 )
296 12: PLD( sfi_pld r1, #-128 )
298 ldmdb \B!, {r7, r8, r10, ip}
299 mov lr, r3, PUSH #\push
302 ldmdb \B!, {r3, r4, r5, r6}
303 orr lr, lr, ip, PULL #\pull
304 mov ip, ip, PUSH #\push
305 orr ip, ip, r10, PULL #\pull
306 mov r10, r10, PUSH #\push
307 orr r10, r10, r8, PULL #\pull
308 mov r8, r8, PUSH #\push
309 orr r8, r8, r7, PULL #\pull
310 mov r7, r7, PUSH #\push
311 orr r7, r7, r6, PULL #\pull
312 mov r6, r6, PUSH #\push
313 orr r6, r6, r5, PULL #\pull
314 mov r5, r5, PUSH #\push
315 orr r5, r5, r4, PULL #\pull
316 mov r4, r4, PUSH #\push
317 orr r4, r4, r3, PULL #\pull
319 stmdb \B!, {r4 - r8, r10, ip, lr}
325 cfi_adjust_cfa_offset (-20)
335 15: mov lr, r3, PUSH #\push
339 orr lr, lr, r3, PULL #\pull
346 16: add r1, r1, #(\pull / 8)
352 backward_copy_shift push=8 pull=24
354 17: backward_copy_shift push=16 pull=16
356 18: backward_copy_shift push=24 pull=8
360 libc_hidden_builtin_def (memmove)