-#ifndef PERL_SEEN_HV_MACRO_H /* compile once */
-#define PERL_SEEN_HV_MACRO_H
+#ifndef PERL_SEEN_HV_MACRO_H_ /* compile once */
+#define PERL_SEEN_HV_MACRO_H_
#if IVSIZE == 8
#define CAN64BITHASH
#endif
+#ifdef CAN64BITHASH
+ #ifndef U64TYPE
+ /* This probably isn't going to work, but failing with a compiler error due to
+ lack of uint64_t is no worse than failing right now with an #error. */
+ #define U64 uint64_t
+ #endif
+#endif
+
+
/*-----------------------------------------------------------------------------
* Endianess and util macros
*
*/
#ifndef U8TO16_LE
- #if (BYTEORDER == 0x1234 || BYTEORDER == 0x12345678)
- #define U8TO16_LE(ptr) ((U32)(ptr)[1]|(U32)(ptr)[0]<<8)
- #define U8TO32_LE(ptr) ((U32)(ptr)[3]|(U32)(ptr)[2]<<8|(U32)(ptr)[1]<<16|(U32)(ptr)[0]<<24)
- #define U8TO64_LE(ptr) ((U64)(ptr)[7]|(U64)(ptr)[6]<<8|(U64)(ptr)[5]<<16|(U64)(ptr)[4]<<24|\
- (U64)(ptr)[3]<<32|(U64)(ptr)[4]<<40|\
- (U64)(ptr)[1]<<48|(U64)(ptr)[0]<<56)
- #elif (BYTEORDER == 0x4321 || BYTEORDER == 0x87654321)
- #define U8TO16_LE(ptr) ((U32)(ptr)[0]|(U32)(ptr)[1]<<8)
- #define U8TO32_LE(ptr) ((U32)(ptr)[0]|(U32)(ptr)[1]<<8|(U32)(ptr)[2]<<16|(U32)(ptr)[3]<<24)
- #define U8TO64_LE(ptr) ((U64)(ptr)[0]|(U64)(ptr)[1]<<8|(U64)(ptr)[2]<<16|(U64)(ptr)[3]<<24|\
- (U64)(ptr)[4]<<32|(U64)(ptr)[5]<<40|\
- (U64)(ptr)[6]<<48|(U64)(ptr)[7]<<56)
- #endif
-#endif
+ #define _shifted_octet(type,ptr,idx,shift) (((type)(((U8*)(ptr))[(idx)]))<<(shift))
+ #ifdef USE_UNALIGNED_PTR_DEREF
+ #define U8TO16_LE(ptr) (*((const U16*)(ptr)))
+ #define U8TO32_LE(ptr) (*((const U32*)(ptr)))
+ #define U8TO64_LE(ptr) (*((const U64*)(ptr)))
+ #else
+ #define U8TO16_LE(ptr) (_shifted_octet(U16,(ptr),0, 0)|\
+ _shifted_octet(U16,(ptr),1, 8))
-#ifdef CAN64BITHASH
- #ifndef U64TYPE
- /* This probably isn't going to work, but failing with a compiler error due to
- lack of uint64_t is no worse than failing right now with an #error. */
- #define U64 uint64_t
- #endif
+ #define U8TO32_LE(ptr) (_shifted_octet(U32,(ptr),0, 0)|\
+ _shifted_octet(U32,(ptr),1, 8)|\
+ _shifted_octet(U32,(ptr),2,16)|\
+ _shifted_octet(U32,(ptr),3,24))
+
+ #define U8TO64_LE(ptr) (_shifted_octet(U64,(ptr),0, 0)|\
+ _shifted_octet(U64,(ptr),1, 8)|\
+ _shifted_octet(U64,(ptr),2,16)|\
+ _shifted_octet(U64,(ptr),3,24)|\
+ _shifted_octet(U64,(ptr),4,32)|\
+ _shifted_octet(U64,(ptr),5,40)|\
+ _shifted_octet(U64,(ptr),6,48)|\
+ _shifted_octet(U64,(ptr),7,56))
+ #endif
#endif
/* Find best way to ROTL32/ROTL64 */