mirror of
https://github.com/torvalds/linux.git
synced 2024-11-17 09:31:50 +00:00
a7f626c194
Original port to early 2.6 kernel using TI COFF toolchain. Brought up to date by Mark Salter <msalter@redhat.com> Signed-off-by: Aurelien Jacquiot <a-jacquiot@ti.com> Signed-off-by: Mark Salter <msalter@redhat.com> Acked-by: Arnd Bergmann <arnd@arndb.de>
171 lines
4.6 KiB
C
171 lines
4.6 KiB
C
/*
|
|
* Port on Texas Instruments TMS320C6x architecture
|
|
*
|
|
* Copyright (C) 2004, 2009, 2010 Texas Instruments Incorporated
|
|
* Author: Aurelien Jacquiot (aurelien.jacquiot@jaluna.com)
|
|
* Rewritten for 2.6.3x: Mark Salter <msalter@redhat.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
#ifndef _ASM_C6X_UNALIGNED_H
|
|
#define _ASM_C6X_UNALIGNED_H
|
|
|
|
#include <linux/swab.h>
|
|
|
|
/*
|
|
* The C64x+ can do unaligned word and dword accesses in hardware
|
|
* using special load/store instructions.
|
|
*/
|
|
|
|
static inline u16 get_unaligned_le16(const void *p)
|
|
{
|
|
const u8 *_p = p;
|
|
return _p[0] | _p[1] << 8;
|
|
}
|
|
|
|
static inline u16 get_unaligned_be16(const void *p)
|
|
{
|
|
const u8 *_p = p;
|
|
return _p[0] << 8 | _p[1];
|
|
}
|
|
|
|
static inline void put_unaligned_le16(u16 val, void *p)
|
|
{
|
|
u8 *_p = p;
|
|
_p[0] = val;
|
|
_p[1] = val >> 8;
|
|
}
|
|
|
|
static inline void put_unaligned_be16(u16 val, void *p)
|
|
{
|
|
u8 *_p = p;
|
|
_p[0] = val >> 8;
|
|
_p[1] = val;
|
|
}
|
|
|
|
static inline u32 get_unaligned32(const void *p)
|
|
{
|
|
u32 val = (u32) p;
|
|
asm (" ldnw .d1t1 *%0,%0\n"
|
|
" nop 4\n"
|
|
: "+a"(val));
|
|
return val;
|
|
}
|
|
|
|
static inline void put_unaligned32(u32 val, void *p)
|
|
{
|
|
asm volatile (" stnw .d2t1 %0,*%1\n"
|
|
: : "a"(val), "b"(p) : "memory");
|
|
}
|
|
|
|
static inline u64 get_unaligned64(const void *p)
|
|
{
|
|
u64 val;
|
|
asm volatile (" ldndw .d1t1 *%1,%0\n"
|
|
" nop 4\n"
|
|
: "=a"(val) : "a"(p));
|
|
return val;
|
|
}
|
|
|
|
static inline void put_unaligned64(u64 val, const void *p)
|
|
{
|
|
asm volatile (" stndw .d2t1 %0,*%1\n"
|
|
: : "a"(val), "b"(p) : "memory");
|
|
}
|
|
|
|
#ifdef CONFIG_CPU_BIG_ENDIAN
|
|
|
|
#define get_unaligned_le32(p) __swab32(get_unaligned32(p))
|
|
#define get_unaligned_le64(p) __swab64(get_unaligned64(p))
|
|
#define get_unaligned_be32(p) get_unaligned32(p)
|
|
#define get_unaligned_be64(p) get_unaligned64(p)
|
|
#define put_unaligned_le32(v, p) put_unaligned32(__swab32(v), (p))
|
|
#define put_unaligned_le64(v, p) put_unaligned64(__swab64(v), (p))
|
|
#define put_unaligned_be32(v, p) put_unaligned32((v), (p))
|
|
#define put_unaligned_be64(v, p) put_unaligned64((v), (p))
|
|
#define get_unaligned __get_unaligned_be
|
|
#define put_unaligned __put_unaligned_be
|
|
|
|
#else
|
|
|
|
#define get_unaligned_le32(p) get_unaligned32(p)
|
|
#define get_unaligned_le64(p) get_unaligned64(p)
|
|
#define get_unaligned_be32(p) __swab32(get_unaligned32(p))
|
|
#define get_unaligned_be64(p) __swab64(get_unaligned64(p))
|
|
#define put_unaligned_le32(v, p) put_unaligned32((v), (p))
|
|
#define put_unaligned_le64(v, p) put_unaligned64((v), (p))
|
|
#define put_unaligned_be32(v, p) put_unaligned32(__swab32(v), (p))
|
|
#define put_unaligned_be64(v, p) put_unaligned64(__swab64(v), (p))
|
|
#define get_unaligned __get_unaligned_le
|
|
#define put_unaligned __put_unaligned_le
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Cause a link-time error if we try an unaligned access other than
|
|
* 1,2,4 or 8 bytes long
|
|
*/
|
|
extern int __bad_unaligned_access_size(void);
|
|
|
|
#define __get_unaligned_le(ptr) (typeof(*(ptr)))({ \
|
|
sizeof(*(ptr)) == 1 ? *(ptr) : \
|
|
(sizeof(*(ptr)) == 2 ? get_unaligned_le16((ptr)) : \
|
|
(sizeof(*(ptr)) == 4 ? get_unaligned_le32((ptr)) : \
|
|
(sizeof(*(ptr)) == 8 ? get_unaligned_le64((ptr)) : \
|
|
__bad_unaligned_access_size()))); \
|
|
})
|
|
|
|
#define __get_unaligned_be(ptr) (__force typeof(*(ptr)))({ \
|
|
sizeof(*(ptr)) == 1 ? *(ptr) : \
|
|
(sizeof(*(ptr)) == 2 ? get_unaligned_be16((ptr)) : \
|
|
(sizeof(*(ptr)) == 4 ? get_unaligned_be32((ptr)) : \
|
|
(sizeof(*(ptr)) == 8 ? get_unaligned_be64((ptr)) : \
|
|
__bad_unaligned_access_size()))); \
|
|
})
|
|
|
|
#define __put_unaligned_le(val, ptr) ({ \
|
|
void *__gu_p = (ptr); \
|
|
switch (sizeof(*(ptr))) { \
|
|
case 1: \
|
|
*(u8 *)__gu_p = (__force u8)(val); \
|
|
break; \
|
|
case 2: \
|
|
put_unaligned_le16((__force u16)(val), __gu_p); \
|
|
break; \
|
|
case 4: \
|
|
put_unaligned_le32((__force u32)(val), __gu_p); \
|
|
break; \
|
|
case 8: \
|
|
put_unaligned_le64((__force u64)(val), __gu_p); \
|
|
break; \
|
|
default: \
|
|
__bad_unaligned_access_size(); \
|
|
break; \
|
|
} \
|
|
(void)0; })
|
|
|
|
#define __put_unaligned_be(val, ptr) ({ \
|
|
void *__gu_p = (ptr); \
|
|
switch (sizeof(*(ptr))) { \
|
|
case 1: \
|
|
*(u8 *)__gu_p = (__force u8)(val); \
|
|
break; \
|
|
case 2: \
|
|
put_unaligned_be16((__force u16)(val), __gu_p); \
|
|
break; \
|
|
case 4: \
|
|
put_unaligned_be32((__force u32)(val), __gu_p); \
|
|
break; \
|
|
case 8: \
|
|
put_unaligned_be64((__force u64)(val), __gu_p); \
|
|
break; \
|
|
default: \
|
|
__bad_unaligned_access_size(); \
|
|
break; \
|
|
} \
|
|
(void)0; })
|
|
|
|
#endif /* _ASM_C6X_UNALIGNED_H */
|