aboutsummaryrefslogtreecommitdiffstats
path: root/arch/mn10300/lib/memmove.S
blob: 20b07b62b77c38bc96bc7a0dc66fef59b98f1f02 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
/* MN10300 Optimised simple memory to memory copy, with support for overlapping
 * regions
 *
 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
 * Written by David Howells (dhowells@redhat.com)
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public Licence
 * as published by the Free Software Foundation; either version
 * 2 of the Licence, or (at your option) any later version.
 */
#include <asm/cache.h>

        .section .text
        .balign	L1_CACHE_BYTES

###############################################################################
#
# void *memmove(void *dst, const void *src, size_t n)
#
###############################################################################
	.globl	memmove
        .type	memmove,@function
memmove:
	# fall back to memcpy if dst < src to work bottom up
	cmp	d1,d0
	bcs	memmove_memcpy

	# work top down
	movm	[d2,d3],(sp)
	mov	d0,(12,sp)
	mov	d1,(16,sp)
	mov	(20,sp),d2			# count
	add	d0,d2,a0			# dst end
	add	d1,d2,a1			# src end
	mov	d0,e3				# the return value

	cmp	+0,d2
	beq	memmove_done			# return if zero-length copy

	# see if the three parameters are all four-byte aligned
	or	d0,d1,d3
	or	d2,d3
	and	+3,d3
	bne	memmove_1			# jump if not

	# we want to transfer as much as we can in chunks of 32 bytes
	add	-4,a1
	cmp	+31,d2
	bls	memmove_4_remainder		# 4-byte aligned remainder

	add	-32,d2
	mov	+32,d3

memmove_4_loop:
	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
	mov	(a1),d1
	sub_sub	+4,a1,+4,a0
	mov	d1,(a0)

	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
	mov	(a1),d1
	sub_sub	+4,a1,+4,a0
	mov	d1,(a0)

	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
	mov	(a1),d1
	sub_sub	+4,a1,+4,a0
	mov	d1,(a0)

	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
	mov	(a1),d1
	sub_sub	+4,a1,+4,a0
	mov	d1,(a0)

	sub	d3,d2
	bcc	memmove_4_loop

	add	d3,d2
	beq	memmove_4_no_remainder

memmove_4_remainder:
	# cut 4-7 words down to 0-3
	cmp	+16,d2
	bcs	memmove_4_three_or_fewer_words
	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
	mov	(a1),d1
	sub_sub	+4,a1,+4,a0
	mov	d1,(a0)
	mov	(a1),e0
	sub_sub	+4,a1,+4,a0
	mov	e0,(a0)
	mov	(a1),e1
	sub_sub	+4,a1,+4,a0
	mov	e1,(a0)
	add	-16,d2
	beq	memmove_4_no_remainder

	# copy the remaining 1, 2 or 3 words
memmove_4_three_or_fewer_words:
	cmp	+8,d2
	bcs	memmove_4_one_word
	beq	memmove_4_two_words
	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
memmove_4_two_words:
	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)
memmove_4_one_word:
	mov	(a1),d0
	sub_sub	+4,a1,+4,a0
	mov	d0,(a0)

memmove_4_no_remainder:
	# check we copied the correct amount
	# TODO: REMOVE CHECK
	sub	e3,a0,d2
	beq	memmove_done
	break
	break
	break

memmove_done:
	mov	e3,a0
	ret	[d2,d3],8

	# handle misaligned copying
memmove_1:
	add	-1,a1
	add	-1,d2
	mov	+1,d3
	setlb					# setlb requires the next insns
						# to occupy exactly 4 bytes

	sub	d3,d2
	movbu	(a1),d0
	sub_sub	d3,a1,d3,a0
	movbu	d0,(a0)
	lcc

	mov	e3,a0
	ret	[d2,d3],8

memmove_memcpy:
	jmp	memcpy

memmove_end:
	.size	memmove, memmove_end-memmove