/* * Copyright (c) 2000-2002,2005 Silicon Graphics, Inc. * All Rights Reserved. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License as * published by the Free Software Foundation. * * This program is distributed in the hope that it would be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ #ifndef __XFS_ARCH_H__ #define __XFS_ARCH_H__ #ifndef XFS_BIG_INUMS # error XFS_BIG_INUMS must be defined true or false #endif #ifdef __KERNEL__ #include <asm/byteorder.h> #ifdef __BIG_ENDIAN #define XFS_NATIVE_HOST 1 #else #undef XFS_NATIVE_HOST #endif #else /* __KERNEL__ */ #if __BYTE_ORDER == __BIG_ENDIAN #define XFS_NATIVE_HOST 1 #else #undef XFS_NATIVE_HOST #endif #ifdef XFS_NATIVE_HOST #define cpu_to_be16(val) ((__force __be16)(__u16)(val)) #define cpu_to_be32(val) ((__force __be32)(__u32)(val)) #define cpu_to_be64(val) ((__force __be64)(__u64)(val)) #define be16_to_cpu(val) ((__force __u16)(__be16)(val)) #define be32_to_cpu(val) ((__force __u32)(__be32)(val)) #define be64_to_cpu(val) ((__force __u64)(__be64)(val)) #else #define cpu_to_be16(val) ((__force __be16)__swab16((__u16)(val))) #define cpu_to_be32(val) ((__force __be32)__swab32((__u32)(val))) #define cpu_to_be64(val) ((__force __be64)__swab64((__u64)(val))) #define be16_to_cpu(val) (__swab16((__force __u16)(__be16)(val))) #define be32_to_cpu(val) (__swab32((__force __u32)(__be32)(val))) #define be64_to_cpu(val) (__swab64((__force __u64)(__be64)(val))) #endif static inline void be16_add_cpu(__be16 *a, __s16 b) { *a = cpu_to_be16(be16_to_cpu(*a) + b); } static inline void be32_add_cpu(__be32 *a, __s32 b) { *a = cpu_to_be32(be32_to_cpu(*a) + b); } static inline void be64_add_cpu(__be64 *a, __s64 b) { *a = cpu_to_be64(be64_to_cpu(*a) + b); } #endif /* __KERNEL__ */ /* * get and set integers from potentially unaligned locations */ #define INT_GET_UNALIGNED_16_BE(pointer) \ ((__u16)((((__u8*)(pointer))[0] << 8) | (((__u8*)(pointer))[1]))) #define INT_SET_UNALIGNED_16_BE(pointer,value) \ { \ ((__u8*)(pointer))[0] = (((value) >> 8) & 0xff); \ ((__u8*)(pointer))[1] = (((value) ) & 0xff); \ } /* * In directories inode numbers are stored as unaligned arrays of unsigned * 8bit integers on disk. * * For v1 directories or v2 directories that contain inode numbers that * do not fit into 32bit the array has eight members, but the first member * is always zero: * * |unused|48-55|40-47|32-39|24-31|16-23| 8-15| 0- 7| * * For v2 directories that only contain entries with inode numbers that fit * into 32bits a four-member array is used: * * |24-31|16-23| 8-15| 0- 7| */ #define XFS_GET_DIR_INO4(di) \ (((__u32)(di).i[0] << 24) | ((di).i[1] << 16) | ((di).i[2] << 8) | ((di).i[3])) #define XFS_PUT_DIR_INO4(from, di) \ do { \ (di).i[0] = (((from) & 0xff000000ULL) >> 24); \ (di).i[1] = (((from) & 0x00ff0000ULL) >> 16); \ (di).i[2] = (((from) & 0x0000ff00ULL) >> 8); \ (di).i[3] = ((from) & 0x000000ffULL); \ } while (0) #define XFS_DI_HI(di) \ (((__u32)(di).i[1] << 16) | ((di).i[2] << 8) | ((di).i[3])) #define XFS_DI_LO(di) \ (((__u32)(di).i[4] << 24) | ((di).i[5] << 16) | ((di).i[6] << 8) | ((di).i[7])) #define XFS_GET_DIR_INO8(di) \ (((xfs_ino_t)XFS_DI_LO(di) & 0xffffffffULL) | \ ((xfs_ino_t)XFS_DI_HI(di) << 32)) #define XFS_PUT_DIR_INO8(from, di) \ do { \ (di).i[0] = 0; \ (di).i[1] = (((from) & 0x00ff000000000000ULL) >> 48); \ (di).i[2] = (((from) & 0x0000ff0000000000ULL) >> 40); \ (di).i[3] = (((from) & 0x000000ff00000000ULL) >> 32); \ (di).i[4] = (((from) & 0x00000000ff000000ULL) >> 24); \ (di).i[5] = (((from) & 0x0000000000ff0000ULL) >> 16); \ (di).i[6] = (((from) & 0x000000000000ff00ULL) >> 8); \ (di).i[7] = ((from) & 0x00000000000000ffULL); \ } while (0) #endif /* __XFS_ARCH_H__ */