mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-12-11 13:56:30 +00:00
2272382879
These are library functions used by ARC700 architecture. Signed-off-by: Alexey Brodkin <abrodkin@synopsys.com> Cc: Vineet Gupta <vgupta@synopsys.com> Cc: Francois Bedard <fbedard@synopsys.com> Cc: Wolfgang Denk <wd@denx.de> Cc: Heiko Schocher <hs@denx.de>
141 lines
2.9 KiB
ArmAsm
141 lines
2.9 KiB
ArmAsm
/*
|
|
* Copyright (C) 2004, 2007-2010, 2011-2014 Synopsys, Inc. All rights reserved.
|
|
*
|
|
* SPDX-License-Identifier: GPL-2.0+
|
|
*/
|
|
|
|
/*
|
|
* ARC700 has a relatively long pipeline and branch prediction, so we want
|
|
* to avoid branches that are hard to predict. On the other hand, the
|
|
* presence of the norm instruction makes it easier to operate on whole
|
|
* words branch-free.
|
|
*/
|
|
|
|
.global strchr
|
|
.align 4
|
|
strchr:
|
|
extb_s %r1, %r1
|
|
asl %r5, %r1, 8
|
|
bmsk %r2, %r0, 1
|
|
or %r5, %r5, %r1
|
|
mov_s %r3, 0x01010101
|
|
breq.d %r2, %r0, .Laligned
|
|
asl %r4, %r5, 16
|
|
sub_s %r0, %r0, %r2
|
|
asl %r7, %r2, 3
|
|
ld_s %r2, [%r0]
|
|
#ifdef __LITTLE_ENDIAN__
|
|
asl %r7, %r3, %r7
|
|
#else /* __BIG_ENDIAN__ */
|
|
lsr %r7, %r3, %r7
|
|
#endif /* _ENDIAN__ */
|
|
or %r5, %r5, %r4
|
|
ror %r4, %r3
|
|
sub %r12, %r2, %r7
|
|
bic_s %r12, %r12, %r2
|
|
and %r12, %r12, %r4
|
|
brne.d %r12, 0, .Lfound0_ua
|
|
xor %r6, %r2, %r5
|
|
ld.a %r2, [%r0, 4]
|
|
sub %r12, %r6, %r7
|
|
bic %r12, %r12, %r6
|
|
#ifdef __LITTLE_ENDIAN__
|
|
and %r7, %r12, %r4
|
|
/* For speed, we want this branch to be unaligned. */
|
|
breq %r7, 0, .Loop
|
|
/* Likewise this one */
|
|
b .Lfound_char
|
|
#else /* __BIG_ENDIAN__ */
|
|
and %r12, %r12, %r4
|
|
/* For speed, we want this branch to be unaligned. */
|
|
breq %r12, 0, .Loop
|
|
lsr_s %r12, %r12, 7
|
|
bic %r2, %r7, %r6
|
|
b.d .Lfound_char_b
|
|
and_s %r2, %r2, %r12
|
|
#endif /* _ENDIAN__ */
|
|
/* We require this code address to be unaligned for speed... */
|
|
.Laligned:
|
|
ld_s %r2, [%r0]
|
|
or %r5, %r5, %r4
|
|
ror %r4, %r3
|
|
/* ... so that this code address is aligned, for itself and ... */
|
|
.Loop:
|
|
sub %r12, %r2, %r3
|
|
bic_s %r12, %r12, %r2
|
|
and %r12, %r12, %r4
|
|
brne.d %r12, 0, .Lfound0
|
|
xor %r6, %r2, %r5
|
|
ld.a %r2, [%r0, 4]
|
|
sub %r12, %r6, %r3
|
|
bic %r12, %r12, %r6
|
|
and %r7, %r12, %r4
|
|
breq %r7, 0, .Loop
|
|
/*
|
|
*... so that this branch is unaligned.
|
|
* Found searched-for character.
|
|
* r0 has already advanced to next word.
|
|
*/
|
|
#ifdef __LITTLE_ENDIAN__
|
|
/*
|
|
* We only need the information about the first matching byte
|
|
* (i.e. the least significant matching byte) to be exact,
|
|
* hence there is no problem with carry effects.
|
|
*/
|
|
.Lfound_char:
|
|
sub %r3, %r7, 1
|
|
bic %r3, %r3, %r7
|
|
norm %r2, %r3
|
|
sub_s %r0, %r0, 1
|
|
asr_s %r2, %r2, 3
|
|
j.d [%blink]
|
|
sub_s %r0, %r0, %r2
|
|
|
|
.balign 4
|
|
.Lfound0_ua:
|
|
mov %r3, %r7
|
|
.Lfound0:
|
|
sub %r3, %r6, %r3
|
|
bic %r3, %r3, %r6
|
|
and %r2, %r3, %r4
|
|
or_s %r12, %r12, %r2
|
|
sub_s %r3, %r12, 1
|
|
bic_s %r3, %r3, %r12
|
|
norm %r3, %r3
|
|
add_s %r0, %r0, 3
|
|
asr_s %r12, %r3, 3
|
|
asl.f 0, %r2, %r3
|
|
sub_s %r0, %r0, %r12
|
|
j_s.d [%blink]
|
|
mov.pl %r0, 0
|
|
#else /* __BIG_ENDIAN__ */
|
|
.Lfound_char:
|
|
lsr %r7, %r7, 7
|
|
|
|
bic %r2, %r7, %r6
|
|
.Lfound_char_b:
|
|
norm %r2, %r2
|
|
sub_s %r0, %r0, 4
|
|
asr_s %r2, %r2, 3
|
|
j.d [%blink]
|
|
add_s %r0, %r0, %r2
|
|
|
|
.Lfound0_ua:
|
|
mov_s %r3, %r7
|
|
.Lfound0:
|
|
asl_s %r2, %r2, 7
|
|
or %r7, %r6, %r4
|
|
bic_s %r12, %r12, %r2
|
|
sub %r2, %r7, %r3
|
|
or %r2, %r2, %r6
|
|
bic %r12, %r2, %r12
|
|
bic.f %r3, %r4, %r12
|
|
norm %r3, %r3
|
|
|
|
add.pl %r3, %r3, 1
|
|
asr_s %r12, %r3, 3
|
|
asl.f 0, %r2, %r3
|
|
add_s %r0, %r0, %r12
|
|
j_s.d [%blink]
|
|
mov.mi %r0, 0
|
|
#endif /* _ENDIAN__ */
|