checksum.S (6298B)
1/* SPDX-License-Identifier: GPL-2.0+ 2 * 3 * $Id: checksum.S,v 1.10 2001/07/06 13:11:32 gniibe Exp $ 4 * 5 * INET An implementation of the TCP/IP protocol suite for the LINUX 6 * operating system. INET is implemented using the BSD Socket 7 * interface as the means of communication with the user level. 8 * 9 * IP/TCP/UDP checksumming routines 10 * 11 * Authors: Jorge Cwik, <jorge@laser.satlink.net> 12 * Arnt Gulbrandsen, <agulbra@nvg.unit.no> 13 * Tom May, <ftom@netcom.com> 14 * Pentium Pro/II routines: 15 * Alexander Kjeldaas <astor@guardian.no> 16 * Finn Arne Gangstad <finnag@guardian.no> 17 * Lots of code moved from tcp.c and ip.c; see those files 18 * for more names. 19 * 20 * Changes: Ingo Molnar, converted csum_partial_copy() to 2.1 exception 21 * handling. 22 * Andi Kleen, add zeroing on error 23 * converted to pure assembler 24 * 25 * SuperH version: Copyright (C) 1999 Niibe Yutaka 26 */ 27 28#include <asm/errno.h> 29#include <linux/linkage.h> 30 31/* 32 * computes a partial checksum, e.g. for TCP/UDP fragments 33 */ 34 35/* 36 * asmlinkage __wsum csum_partial(const void *buf, int len, __wsum sum); 37 */ 38 39.text 40ENTRY(csum_partial) 41 /* 42 * Experiments with Ethernet and SLIP connections show that buff 43 * is aligned on either a 2-byte or 4-byte boundary. We get at 44 * least a twofold speedup on 486 and Pentium if it is 4-byte aligned. 45 * Fortunately, it is easy to convert 2-byte alignment to 4-byte 46 * alignment for the unrolled loop. 47 */ 48 mov r4, r0 49 tst #3, r0 ! Check alignment. 50 bt/s 2f ! Jump if alignment is ok. 51 mov r4, r7 ! Keep a copy to check for alignment 52 ! 53 tst #1, r0 ! Check alignment. 54 bt 21f ! Jump if alignment is boundary of 2bytes. 55 56 ! buf is odd 57 tst r5, r5 58 add #-1, r5 59 bt 9f 60 mov.b @r4+, r0 61 extu.b r0, r0 62 addc r0, r6 ! t=0 from previous tst 63 mov r6, r0 64 shll8 r6 65 shlr16 r0 66 shlr8 r0 67 or r0, r6 68 mov r4, r0 69 tst #2, r0 70 bt 2f 7121: 72 ! buf is 2 byte aligned (len could be 0) 73 add #-2, r5 ! Alignment uses up two bytes. 74 cmp/pz r5 ! 75 bt/s 1f ! Jump if we had at least two bytes. 76 clrt 77 bra 6f 78 add #2, r5 ! r5 was < 2. Deal with it. 791: 80 mov.w @r4+, r0 81 extu.w r0, r0 82 addc r0, r6 83 bf 2f 84 add #1, r6 852: 86 ! buf is 4 byte aligned (len could be 0) 87 mov r5, r1 88 mov #-5, r0 89 shld r0, r1 90 tst r1, r1 91 bt/s 4f ! if it's =0, go to 4f 92 clrt 93 .align 2 943: 95 mov.l @r4+, r0 96 mov.l @r4+, r2 97 mov.l @r4+, r3 98 addc r0, r6 99 mov.l @r4+, r0 100 addc r2, r6 101 mov.l @r4+, r2 102 addc r3, r6 103 mov.l @r4+, r3 104 addc r0, r6 105 mov.l @r4+, r0 106 addc r2, r6 107 mov.l @r4+, r2 108 addc r3, r6 109 addc r0, r6 110 addc r2, r6 111 movt r0 112 dt r1 113 bf/s 3b 114 cmp/eq #1, r0 115 ! here, we know r1==0 116 addc r1, r6 ! add carry to r6 1174: 118 mov r5, r0 119 and #0x1c, r0 120 tst r0, r0 121 bt 6f 122 ! 4 bytes or more remaining 123 mov r0, r1 124 shlr2 r1 125 mov #0, r2 1265: 127 addc r2, r6 128 mov.l @r4+, r2 129 movt r0 130 dt r1 131 bf/s 5b 132 cmp/eq #1, r0 133 addc r2, r6 134 addc r1, r6 ! r1==0 here, so it means add carry-bit 1356: 136 ! 3 bytes or less remaining 137 mov #3, r0 138 and r0, r5 139 tst r5, r5 140 bt 9f ! if it's =0 go to 9f 141 mov #2, r1 142 cmp/hs r1, r5 143 bf 7f 144 mov.w @r4+, r0 145 extu.w r0, r0 146 cmp/eq r1, r5 147 bt/s 8f 148 clrt 149 shll16 r0 150 addc r0, r6 1517: 152 mov.b @r4+, r0 153 extu.b r0, r0 154#ifndef __LITTLE_ENDIAN__ 155 shll8 r0 156#endif 1578: 158 addc r0, r6 159 mov #0, r0 160 addc r0, r6 1619: 162 ! Check if the buffer was misaligned, if so realign sum 163 mov r7, r0 164 tst #1, r0 165 bt 10f 166 mov r6, r0 167 shll8 r6 168 shlr16 r0 169 shlr8 r0 170 or r0, r6 17110: 172 rts 173 mov r6, r0 174 175/* 176unsigned int csum_partial_copy_generic (const char *src, char *dst, int len) 177 */ 178 179/* 180 * Copy from ds while checksumming, otherwise like csum_partial with initial 181 * sum being ~0U 182 */ 183 184#define EXC(...) \ 185 9999: __VA_ARGS__ ; \ 186 .section __ex_table, "a"; \ 187 .long 9999b, 6001f ; \ 188 .previous 189 190! 191! r4: const char *SRC 192! r5: char *DST 193! r6: int LEN 194! 195ENTRY(csum_partial_copy_generic) 196 mov #-1,r7 197 mov #3,r0 ! Check src and dest are equally aligned 198 mov r4,r1 199 and r0,r1 200 and r5,r0 201 cmp/eq r1,r0 202 bf 3f ! Different alignments, use slow version 203 tst #1,r0 ! Check dest word aligned 204 bf 3f ! If not, do it the slow way 205 206 mov #2,r0 207 tst r0,r5 ! Check dest alignment. 208 bt 2f ! Jump if alignment is ok. 209 add #-2,r6 ! Alignment uses up two bytes. 210 cmp/pz r6 ! Jump if we had at least two bytes. 211 bt/s 1f 212 clrt 213 add #2,r6 ! r6 was < 2. Deal with it. 214 bra 4f 215 mov r6,r2 216 2173: ! Handle different src and dest alignments. 218 ! This is not common, so simple byte by byte copy will do. 219 mov r6,r2 220 shlr r6 221 tst r6,r6 222 bt 4f 223 clrt 224 .align 2 2255: 226EXC( mov.b @r4+,r1 ) 227EXC( mov.b @r4+,r0 ) 228 extu.b r1,r1 229EXC( mov.b r1,@r5 ) 230EXC( mov.b r0,@(1,r5) ) 231 extu.b r0,r0 232 add #2,r5 233 234#ifdef __LITTLE_ENDIAN__ 235 shll8 r0 236#else 237 shll8 r1 238#endif 239 or r1,r0 240 241 addc r0,r7 242 movt r0 243 dt r6 244 bf/s 5b 245 cmp/eq #1,r0 246 mov #0,r0 247 addc r0, r7 248 249 mov r2, r0 250 tst #1, r0 251 bt 7f 252 bra 5f 253 clrt 254 255 ! src and dest equally aligned, but to a two byte boundary. 256 ! Handle first two bytes as a special case 257 .align 2 2581: 259EXC( mov.w @r4+,r0 ) 260EXC( mov.w r0,@r5 ) 261 add #2,r5 262 extu.w r0,r0 263 addc r0,r7 264 mov #0,r0 265 addc r0,r7 2662: 267 mov r6,r2 268 mov #-5,r0 269 shld r0,r6 270 tst r6,r6 271 bt/s 2f 272 clrt 273 .align 2 2741: 275EXC( mov.l @r4+,r0 ) 276EXC( mov.l @r4+,r1 ) 277 addc r0,r7 278EXC( mov.l r0,@r5 ) 279EXC( mov.l r1,@(4,r5) ) 280 addc r1,r7 281 282EXC( mov.l @r4+,r0 ) 283EXC( mov.l @r4+,r1 ) 284 addc r0,r7 285EXC( mov.l r0,@(8,r5) ) 286EXC( mov.l r1,@(12,r5) ) 287 addc r1,r7 288 289EXC( mov.l @r4+,r0 ) 290EXC( mov.l @r4+,r1 ) 291 addc r0,r7 292EXC( mov.l r0,@(16,r5) ) 293EXC( mov.l r1,@(20,r5) ) 294 addc r1,r7 295 296EXC( mov.l @r4+,r0 ) 297EXC( mov.l @r4+,r1 ) 298 addc r0,r7 299EXC( mov.l r0,@(24,r5) ) 300EXC( mov.l r1,@(28,r5) ) 301 addc r1,r7 302 add #32,r5 303 movt r0 304 dt r6 305 bf/s 1b 306 cmp/eq #1,r0 307 mov #0,r0 308 addc r0,r7 309 3102: mov r2,r6 311 mov #0x1c,r0 312 and r0,r6 313 cmp/pl r6 314 bf/s 4f 315 clrt 316 shlr2 r6 3173: 318EXC( mov.l @r4+,r0 ) 319 addc r0,r7 320EXC( mov.l r0,@r5 ) 321 add #4,r5 322 movt r0 323 dt r6 324 bf/s 3b 325 cmp/eq #1,r0 326 mov #0,r0 327 addc r0,r7 3284: mov r2,r6 329 mov #3,r0 330 and r0,r6 331 cmp/pl r6 332 bf 7f 333 mov #2,r1 334 cmp/hs r1,r6 335 bf 5f 336EXC( mov.w @r4+,r0 ) 337EXC( mov.w r0,@r5 ) 338 extu.w r0,r0 339 add #2,r5 340 cmp/eq r1,r6 341 bt/s 6f 342 clrt 343 shll16 r0 344 addc r0,r7 3455: 346EXC( mov.b @r4+,r0 ) 347EXC( mov.b r0,@r5 ) 348 extu.b r0,r0 349#ifndef __LITTLE_ENDIAN__ 350 shll8 r0 351#endif 3526: addc r0,r7 353 mov #0,r0 354 addc r0,r7 3557: 356 357# Exception handler: 358.section .fixup, "ax" 359 3606001: 361 rts 362 mov #0,r0 363.previous 364 rts 365 mov r7,r0