Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 85d20dee authored by Al Viro's avatar Al Viro Committed by David S. Miller
Browse files

[NET]: M32R checksum annotations and cleanups.



* sanitize prototypes, annotate
* ntohs -> shift in checksum calculations in l-e case
* kill shift-by-16 in checksum calculations

Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 32252996
Loading
Loading
Loading
Loading
+5 −7
Original line number Diff line number Diff line
@@ -27,9 +27,8 @@
/*
 * Copy while checksumming, otherwise like csum_partial
 */
unsigned int
csum_partial_copy_nocheck (const unsigned char *src, unsigned char *dst,
                           int len, unsigned int sum)
__wsum
csum_partial_copy_nocheck (const void *src, void *dst, int len, __wsum sum)
{
	sum = csum_partial(src, len, sum);
	memcpy(dst, src, len);
@@ -42,10 +41,9 @@ EXPORT_SYMBOL(csum_partial_copy_nocheck);
 * Copy from userspace and compute checksum.  If we catch an exception
 * then zero the rest of the buffer.
 */
unsigned int
csum_partial_copy_from_user (const unsigned char __user *src,
			     unsigned char *dst,
			     int len, unsigned int sum, int *err_ptr)
__wsum
csum_partial_copy_from_user (const void __user *src, void *dst,
			     int len, __wsum sum, int *err_ptr)
{
	int missing;

+24 −28
Original line number Diff line number Diff line
@@ -31,8 +31,7 @@
 *
 * it's best to have buff aligned on a 32-bit boundary
 */
asmlinkage unsigned int csum_partial(const unsigned char *buff,
				     int len, unsigned int sum);
asmlinkage __wsum csum_partial(const void *buff, int len, __wsum sum);

/*
 * The same as csum_partial, but copies from src while it checksums.
@@ -40,24 +39,22 @@ asmlinkage unsigned int csum_partial(const unsigned char *buff,
 * Here even more important to align src and dst on a 32-bit (or even
 * better 64-bit) boundary
 */
extern unsigned int csum_partial_copy_nocheck(const unsigned char *src,
					      unsigned char *dst,
                                              int len, unsigned int sum);
extern __wsum csum_partial_copy_nocheck(const void *src, void *dst,
                                              int len, __wsum sum);

/*
 * This is a new version of the above that records errors it finds in *errp,
 * but continues and zeros thre rest of the buffer.
 */
extern unsigned int csum_partial_copy_from_user(const unsigned char __user *src,
                                                unsigned char *dst,
                                                int len, unsigned int sum,
extern __wsum csum_partial_copy_from_user(const void __user *src, void *dst,
                                                int len, __wsum sum,
                                                int *err_ptr);

/*
 *	Fold a partial checksum
 */

static inline unsigned int csum_fold(unsigned int sum)
static inline __sum16 csum_fold(__wsum sum)
{
	unsigned long tmpreg;
	__asm__(
@@ -72,16 +69,17 @@ static inline unsigned int csum_fold(unsigned int sum)
		: "0"  (sum)
		: "cbit"
	);
	return sum;
	return (__force __sum16)sum;
}

/*
 * This is a version of ip_compute_csum() optimized for IP headers,
 * which always checksum on 4 octet boundaries.
 */
static inline unsigned short ip_fast_csum(unsigned char * iph,
					  unsigned int ihl) {
	unsigned long sum, tmpreg0, tmpreg1;
static inline __sum16 ip_fast_csum(const void *iph, unsigned int ihl)
{
	unsigned long tmpreg0, tmpreg1;
	__wsum sum;

	__asm__ __volatile__(
		"	ld	%0, @%1+ \n"
@@ -115,16 +113,15 @@ static inline unsigned short ip_fast_csum(unsigned char * iph,
	return csum_fold(sum);
}

static inline unsigned long csum_tcpudp_nofold(unsigned long saddr,
					       unsigned long daddr,
static inline __wsum csum_tcpudp_nofold(__be32 saddr, __be32 daddr,
					       unsigned short len,
					       unsigned short proto,
					       unsigned int sum)
					       __wsum sum)
{
#if defined(__LITTLE_ENDIAN)
	unsigned long len_proto = (ntohs(len)<<16)+proto*256;
	unsigned long len_proto = (proto + len) << 8;
#else
	unsigned long len_proto = (proto<<16)+len;
	unsigned long len_proto = proto + len;
#endif
	unsigned long tmpreg;

@@ -147,11 +144,10 @@ static inline unsigned long csum_tcpudp_nofold(unsigned long saddr,
 * computes the checksum of the TCP/UDP pseudo-header
 * returns a 16-bit checksum, already complemented
 */
static inline unsigned short int csum_tcpudp_magic(unsigned long saddr,
						   unsigned long daddr,
static inline __sum16 csum_tcpudp_magic(__be32 saddr, __be32 daddr,
						   unsigned short len,
						   unsigned short proto,
						   unsigned int sum)
						   __wsum sum)
{
	return csum_fold(csum_tcpudp_nofold(saddr,daddr,len,proto,sum));
}
@@ -161,16 +157,16 @@ static inline unsigned short int csum_tcpudp_magic(unsigned long saddr,
 * in icmp.c
 */

static inline unsigned short ip_compute_csum(unsigned char * buff, int len) {
static inline __sum16 ip_compute_csum(const void *buff, int len)
{
	return csum_fold (csum_partial(buff, len, 0));
}

#define _HAVE_ARCH_IPV6_CSUM
static inline unsigned short int csum_ipv6_magic(struct in6_addr *saddr,
						 struct in6_addr *daddr,
						 __u16 len,
						 unsigned short proto,
						 unsigned int sum)
static inline __sum16 csum_ipv6_magic(const struct in6_addr *saddr,
				      const struct in6_addr *daddr,
				      __u32 len, unsigned short proto,
				      __wsum sum)
{
	unsigned long tmpreg0, tmpreg1, tmpreg2, tmpreg3;
	__asm__(
@@ -197,7 +193,7 @@ static inline unsigned short int csum_ipv6_magic(struct in6_addr *saddr,
		: "=&r" (sum), "=&r" (tmpreg0), "=&r" (tmpreg1),
		  "=&r" (tmpreg2), "=&r" (tmpreg3)
		: "r" (saddr), "r" (daddr),
		  "r" (htonl((__u32) (len))), "r" (htonl(proto)), "0" (sum)
		  "r" (htonl(len)), "r" (htonl(proto)), "0" (sum)
		: "cbit"
	);