187 lines
6.2 KiB
ArmAsm
187 lines
6.2 KiB
ArmAsm
|
/* $NetBSD: memset_neon.S,v 1.1 2012/12/13 01:41:59 matt Exp $ */
|
||
|
|
||
|
/*-
|
||
|
* Copyright (c) 2012 The NetBSD Foundation, Inc.
|
||
|
* All rights reserved.
|
||
|
*
|
||
|
* This code is derived from software contributed to The NetBSD Foundation
|
||
|
* by Matt Thomas of 3am Software Foundry.
|
||
|
*
|
||
|
* Redistribution and use in source and binary forms, with or without
|
||
|
* modification, are permitted provided that the following conditions
|
||
|
* are met:
|
||
|
* 1. Redistributions of source code must retain the above copyright
|
||
|
* notice, this list of conditions and the following disclaimer.
|
||
|
* 2. Redistributions in binary form must reproduce the above copyright
|
||
|
* notice, this list of conditions and the following disclaimer in the
|
||
|
* documentation and/or other materials provided with the distribution.
|
||
|
*
|
||
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
||
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
||
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
||
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
||
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
||
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||
|
* POSSIBILITY OF SUCH DAMAGE.
|
||
|
*/
|
||
|
#include <machine/asm.h>
|
||
|
#include "assym.h"
|
||
|
|
||
|
/*
|
||
|
* memset: Sets a block of memory to the specified value
|
||
|
* Using NEON instructions
|
||
|
*
|
||
|
* On entry:
|
||
|
* r0 - dest address
|
||
|
* r1 - byte to write
|
||
|
* r2 - number of bytes to write
|
||
|
*
|
||
|
* On exit:
|
||
|
* r0 - dest address
|
||
|
*/
|
||
|
/* LINTSTUB: Func: void *memset(void *, int, size_t) */
|
||
|
ENTRY(memset)
|
||
|
and r3, r1, #0xff /* We deal with bytes */
|
||
|
mov r1, r2
|
||
|
mov ip, r0 /* r0 needs to stay the same */
|
||
|
|
||
|
vdup.8 q0, r3 /* move fill to SIMD */
|
||
|
/* we no longer need to keep the fill value in a ARM register */
|
||
|
|
||
|
/* Ok first we will dword align the address */
|
||
|
ands r2, ip, #7 /* grab the bottom three bits */
|
||
|
beq .Lmemset_dwordaligned /* The addr is dword aligned */
|
||
|
|
||
|
bic ip, ip, #7 /* clear bottom three bits of addr */
|
||
|
vldr d7, [ip] /* load from memory */
|
||
|
add r1, r1, r2 /* add "pre-fill" to length */
|
||
|
lsl r2, r2, #3 /* byte to no-fill bit count */
|
||
|
|
||
|
#ifdef __ARMEB__
|
||
|
neg r1, r1 /* start from the MSB */
|
||
|
#endif
|
||
|
vmov s4, r2 /* move to SIMD d2 */
|
||
|
vmvn.u64 d3, #0 /* set all ones */
|
||
|
vshl.u64 d3, d3, d2 /* create a no-fill mask */
|
||
|
vmvn.u64 d2, d3 /* invert mask for a fill-mask */
|
||
|
vand d7, d7, d3 /* mask out fill bits */
|
||
|
vand d2, d0, d2 /* mask out no-fill bits */
|
||
|
vorr d7, d2, d7 /* merge fill and memory */
|
||
|
|
||
|
cmp r1, #8 /* Do we have less than 8 bytes */
|
||
|
movlt r2, #0 /* indicate this is the last word */
|
||
|
blt .Lmemset_lessthaneight_noload
|
||
|
|
||
|
vstmia ip!, {d7} /* write back to memory */
|
||
|
subs r1, r1, #8 /* and remove 8 bytes from the length */
|
||
|
RETc(eq)
|
||
|
|
||
|
/* We are now doubleword aligned */
|
||
|
.Lmemset_dwordaligned:
|
||
|
vmov q1, q0 /* put fill in q1 (d2-d3) */
|
||
|
vmov q2, q0 /* put fill in q2 (d4-d5) */
|
||
|
vmov q3, q0 /* put fill in q3 (d6-d7) */
|
||
|
|
||
|
and r2, ip, #63 /* check for 64-byte alignment */
|
||
|
beq .Lmemset_8dwordaligned
|
||
|
/*
|
||
|
* Let's align to a 64-byte boundary so that stores don't cross
|
||
|
* cacheline boundaries. We also know we have at least 128-bytes to
|
||
|
* copy so we don't have to worry about the length at the moment.
|
||
|
*/
|
||
|
rsb r2, r2, #64 /* how many bytes until 64 bytes */
|
||
|
cmp r1, r2 /* compare against length */
|
||
|
andlt r2, r1, #0x38 /* if < len, use trunc(len, 8) */
|
||
|
subs r1, r1, r2 /* subtract from len */
|
||
|
add pc, pc, r2 /* and jump to it */
|
||
|
nop
|
||
|
RETc(eq); b .Lmemset_lessthaneight
|
||
|
vstmia ip!, {d0}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d1}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d2}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d3}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d4}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d5}; b .Lmemset_8dwordaligned
|
||
|
vstmia ip!, {d0-d6}
|
||
|
.Lmemset_8dwordaligned:
|
||
|
vmov d0, d1 /* restore in case of unaligned start */
|
||
|
cmp r1, #8 /* do we have less than 8 bytes */
|
||
|
movlt r2, #0 /* indicate last word */
|
||
|
blt .Lmemset_lessthaneight
|
||
|
|
||
|
cmp r1, #512
|
||
|
blt .Lmemset_sub512
|
||
|
|
||
|
/* Do 512 bytes at a time */
|
||
|
mov r2, #512
|
||
|
.Lmemset_512:
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
vstmia ip!, {d0-d7}
|
||
|
.Lmemset_0:
|
||
|
subs r1, r1, r2
|
||
|
RETc(eq) /* return if done */
|
||
|
cmp r1, #512
|
||
|
bge .Lmemset_512
|
||
|
|
||
|
/*
|
||
|
* We have less than 512 bytes left, but since the sequence above
|
||
|
* store 64 bytes at a time, we determine the number of instructions
|
||
|
* we need to store the remainder (if >= 64 bytes) and execute that
|
||
|
* many vstmia.
|
||
|
*/
|
||
|
.Lmemset_sub512:
|
||
|
lsr r2, r1, #6 /* divide by 64 */
|
||
|
lslne r4, r2, #2 /* multiply by 4 */
|
||
|
addne r4, r4, #1f + 8 - .Lmemset_0
|
||
|
/* add the # of bytes between */
|
||
|
1: subne pc, r4 /* and go */
|
||
|
|
||
|
/*
|
||
|
* We have less than 64 bytes to copy on a 8dword aligned address
|
||
|
*/
|
||
|
and r2, r1, #56 /* get # of full dwords */
|
||
|
ands r1, r1, #7 /* get # of extra bytes */
|
||
|
beq .Lmemset_finalstore
|
||
|
/*
|
||
|
* The last word is a partial fill so load its value and update it
|
||
|
* to include the fill value.
|
||
|
*/
|
||
|
.Lmemset_lessthaneight:
|
||
|
vldr d7, [ip, r2] /* load the last partial dword */
|
||
|
.Lmemset_lessthaneight_noload:
|
||
|
lsl r1, r1, #3 /* byte to fill bit count */
|
||
|
#ifdef __ARMEB__
|
||
|
neg r1, r1 /* start from the MSB */
|
||
|
#endif
|
||
|
vmov s4, r1 /* move to SIMD d2 */
|
||
|
vmvn.u64 d3, #0 /* set all ones */
|
||
|
vshl.u64 d3, d3, d2 /* create a no-fill mask */
|
||
|
vmvn.u64 d2, d3 /* invert mask */
|
||
|
vand d7, d7, d2 /* keep no-fill bits */
|
||
|
vand d2, d0, d3 /* mask out no-fill bits */
|
||
|
vorr d7, d2, d7 /* merge fill and no-fill */
|
||
|
vmov q1, q0 /* restore d2 & d3 */
|
||
|
add r2, r2, #8 /* compensate for the partial dword */
|
||
|
.Lmemset_finalstore:
|
||
|
add pc, pc, r2 /* and jump to it */
|
||
|
nop
|
||
|
vstr d7, [ip]; RET
|
||
|
vstmia ip, {d6-d7}; RET
|
||
|
vstmia ip, {d5-d7}; RET
|
||
|
vstmia ip, {d4-d7}; RET
|
||
|
vstmia ip, {d3-d7}; RET
|
||
|
vstmia ip, {d2-d7}; RET
|
||
|
vstmia ip, {d1-d7}; RET
|
||
|
vstmia ip, {d0-d7}; RET
|
||
|
END(memset)
|