patch-2.3.16 linux/include/asm-sh/string.h

Next file: linux/include/asm-sh/system.h
Previous file: linux/include/asm-sh/statfs.h
Back to the patch index
Back to the overall index

diff -u --recursive --new-file v2.3.15/linux/include/asm-sh/string.h linux/include/asm-sh/string.h
@@ -0,0 +1,140 @@
+#ifndef __ASM_SH_STRING_H
+#define __ASM_SH_STRING_H
+
+/*
+ * Copyright (C) 1999 Niibe Yutaka
+ * But consider these trivial functions to be public domain.
+ */
+
+#define __HAVE_ARCH_STRCPY
+extern __inline__ char *strcpy(char *__dest, const char *__src)
+{
+	register char *__xdest = __dest;
+	unsigned long __dummy;
+
+	__asm__ __volatile__("1:\n\t"
+			     "mov.b	@%1+,%2\n\t"
+			     "mov.b	%2,@%0\n\t"
+			     "cmp/eq	#0,%2\n\t"
+			     "bf/s	1b\n\t"
+			     " add	#1,%0\n\t"
+			     :"=&r" (__dest), "=&r" (__src), "=&z" (__dummy)
+			     :"0" (__dest), "1" (__src)
+			     :"memory");
+
+	return __xdest;
+}
+
+#define __HAVE_ARCH_STRNCPY
+extern __inline__ char *strncpy(char *__dest, const char *__src, size_t __n)
+{
+	register char *__xdest = __dest;
+	unsigned long __dummy;
+
+	if (__n == 0)
+		return __xdest;
+
+	__asm__ __volatile__(
+		"1:\n"
+		"mov.b	@%1+,%2\n\t"
+		"mov.b  %2,@%0\n\t"
+		"cmp/eq #0,%2\n\t"
+		"bt/s   2f\n\t"
+		" cmp/eq %5,%1\n\t"
+		"bf/s   1b\n\t"
+		" add    #1,%0\n"
+		"2:"
+		: "=r" (__dest), "=r" (__src), "=&z" (__dummy)
+		: "0" (__dest), "1" (__src), "r" (__src+__n)
+		: "memory");
+
+	return __xdest;
+}
+
+#define __HAVE_ARCH_STRCMP
+extern __inline__ int strcmp(const char *__cs, const char *__ct)
+{
+	register int __res;
+	unsigned long __dummy;
+
+	__asm__ __volatile__(
+		"mov.b	@%1+,%3\n"
+		"1:\n\t"
+		"mov.b	@%0+,%2\n\t"
+		"cmp/eq #0,%3\n\t"
+		"bt	2f\n\t"
+		"cmp/eq %2,%3\n\t"
+		"bt/s	1b\n\t"
+		" mov.b	@%1+,%3\n\t"
+		"add	#-2,%1\n\t"
+		"mov.b	@%1,%3\n\t"
+		"sub	%3,%2\n"
+		"2:"
+		: "=r" (__cs), "=r" (__ct), "=&r" (__res), "=&z" (__dummy)
+		: "0" (__cs), "1" (__ct));
+
+	return __res;
+}
+
+#define __HAVE_ARCH_STRNCMP
+extern __inline__ int strncmp(const char *__cs, const char *__ct, size_t __n)
+{
+	register int __res;
+	unsigned long __dummy;
+
+	__asm__ __volatile__(
+		"mov.b	@%1+,%3\n"
+		"1:\n\t"
+		"mov.b	@%0+,%2\n\t"
+		"cmp/eq %6,%0\n\t"
+		"bt/s	2f\n\t"
+		" cmp/eq #0,%3\n\t"
+		"bt/s	3f\n\t"
+		" cmp/eq %3,%2\n\t"
+		"bt/s	1b\n\t"
+		" mov.b	@%1+,%3\n\t"
+		"add	#-2,%1\n\t"
+		"mov.b	@%1,%3\n"
+		"2:\n\t"
+		"sub	%3,%2\n"
+		"3:"
+		:"=r" (__cs), "=r" (__ct), "=&r" (__res), "=&z" (__dummy)
+		: "0" (__cs), "1" (__ct), "r" (__cs+__n));
+
+	return __res;
+}
+
+#define __HAVE_ARCH_MEMSET
+extern void *memset(void *__s, int __c, size_t __count);
+
+#define __HAVE_ARCH_MEMCPY
+extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
+
+#define __HAVE_ARCH_MEMMOVE
+extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
+
+/* Don't build bcopy at all ...  */
+#define __HAVE_ARCH_BCOPY
+
+#define __HAVE_ARCH_MEMSCAN
+extern __inline__ void *memscan(void *__addr, int __c, size_t __size)
+{
+	register char *__end;
+	unsigned long __dummy;
+
+	__asm__("mov	#0,%2\n"
+		"1:\n\t"
+		"cmp/eq	%0,%1\n\t"
+		"bt	2f\n\t"
+		"cmp/eq	%5,%2\n\t"
+		"bf/s	1b\n\t"
+		" mov.b	@%0+,%2\n\t"
+		"add	#-1,%0\n\t"
+		"2:"
+		: "=r" (__addr), "=r" (__end), "=&z" (__dummy)
+		: "0" (__addr), "1" ((char *)__addr + __size), "r" (__c));
+
+	return __addr;
+}
+
+#endif /* __ASM_SH_STRING_H */

FUNET's LINUX-ADM group, linux-adm@nic.funet.fi
TCL-scripts by Sam Shen (who was at: slshen@lbl.gov)