r282569 - [Headers] Replace stray indentation with tabs with spaces. NFC.

Martin Storsjo via cfe-commits cfe-commits at lists.llvm.org
Wed Sep 28 02:34:51 PDT 2016


Author: mstorsjo
Date: Wed Sep 28 04:34:51 2016
New Revision: 282569

URL: http://llvm.org/viewvc/llvm-project?rev=282569&view=rev
Log:
[Headers] Replace stray indentation with tabs with spaces. NFC.

This matches the rest of the surrounding file.

Modified:
    cfe/trunk/lib/Headers/intrin.h

Modified: cfe/trunk/lib/Headers/intrin.h
URL: http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/intrin.h?rev=282569&r1=282568&r2=282569&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/intrin.h (original)
+++ cfe/trunk/lib/Headers/intrin.h Wed Sep 28 04:34:51 2016
@@ -595,7 +595,7 @@ _InterlockedExchangeAdd_rel(long volatil
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedExchangeAdd64(__int64 volatile *_Addend, __int64 _Value) {
-	return __atomic_fetch_add(_Addend, _Value, __ATOMIC_SEQ_CST);
+  return __atomic_fetch_add(_Addend, _Value, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -618,7 +618,7 @@ _InterlockedExchangeAdd64_rel(__int64 vo
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedExchangeSub64(__int64 volatile *_Subend, __int64 _Value) {
-	return __atomic_fetch_sub(_Subend, _Value, __ATOMIC_SEQ_CST);
+  return __atomic_fetch_sub(_Subend, _Value, __ATOMIC_SEQ_CST);
 }
 #endif
 /*----------------------------------------------------------------------------*\
@@ -653,7 +653,7 @@ _InterlockedIncrement_rel(long volatile
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedIncrement64(__int64 volatile *_Value) {
-	return __atomic_add_fetch(_Value, 1, __ATOMIC_SEQ_CST);
+  return __atomic_add_fetch(_Value, 1, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -702,7 +702,7 @@ _InterlockedDecrement_rel(long volatile
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedDecrement64(__int64 volatile *_Value) {
-	return __atomic_sub_fetch(_Value, 1, __ATOMIC_SEQ_CST);
+  return __atomic_sub_fetch(_Value, 1, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -763,7 +763,7 @@ _InterlockedAnd_rel(long volatile *_Valu
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedAnd64(__int64 volatile *_Value, __int64 _Mask) {
-	return __atomic_fetch_and(_Value, _Mask, __ATOMIC_SEQ_CST);
+  return __atomic_fetch_and(_Value, _Mask, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -824,7 +824,7 @@ _InterlockedOr_rel(long volatile *_Value
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedOr64(__int64 volatile *_Value, __int64 _Mask) {
-	return __atomic_fetch_or(_Value, _Mask, __ATOMIC_SEQ_CST);
+  return __atomic_fetch_or(_Value, _Mask, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -885,7 +885,7 @@ _InterlockedXor_rel(long volatile *_Valu
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedXor64(__int64 volatile *_Value, __int64 _Mask) {
-	return __atomic_fetch_xor(_Value, _Mask, __ATOMIC_SEQ_CST);
+  return __atomic_fetch_xor(_Value, _Mask, __ATOMIC_SEQ_CST);
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)
@@ -955,8 +955,8 @@ _InterlockedExchange_rel(long volatile *
 #if defined(__x86_64__) || defined(__arm__) || defined(__aarch64__)
 static __inline__ __int64 __DEFAULT_FN_ATTRS
 _InterlockedExchange64(__int64 volatile *_Target, __int64 _Value) {
-	__atomic_exchange(_Target, &_Value, &_Value, __ATOMIC_SEQ_CST);
-	return _Value;
+  __atomic_exchange(_Target, &_Value, &_Value, __ATOMIC_SEQ_CST);
+  return _Value;
 }
 #endif
 #if defined(__arm__) || defined(__aarch64__)




More information about the cfe-commits mailing list