| /**************************************************************************** |
| ** |
| ** Copyright (C) 2011 Nokia Corporation and/or its subsidiary(-ies). |
| ** All rights reserved. |
| ** Contact: Nokia Corporation (qt-info@nokia.com) |
| ** |
| ** This file is part of the QtCore module of the Qt Toolkit. |
| ** |
| ** $QT_BEGIN_LICENSE:LGPL$ |
| ** GNU Lesser General Public License Usage |
| ** This file may be used under the terms of the GNU Lesser General Public |
| ** License version 2.1 as published by the Free Software Foundation and |
| ** appearing in the file LICENSE.LGPL included in the packaging of this |
| ** file. Please review the following information to ensure the GNU Lesser |
| ** General Public License version 2.1 requirements will be met: |
| ** http://www.gnu.org/licenses/old-licenses/lgpl-2.1.html. |
| ** |
| ** In addition, as a special exception, Nokia gives you certain additional |
| ** rights. These rights are described in the Nokia Qt LGPL Exception |
| ** version 1.1, included in the file LGPL_EXCEPTION.txt in this package. |
| ** |
| ** GNU General Public License Usage |
| ** Alternatively, this file may be used under the terms of the GNU General |
| ** Public License version 3.0 as published by the Free Software Foundation |
| ** and appearing in the file LICENSE.GPL included in the packaging of this |
| ** file. Please review the following information to ensure the GNU General |
| ** Public License version 3.0 requirements will be met: |
| ** http://www.gnu.org/copyleft/gpl.html. |
| ** |
| ** Other Usage |
| ** Alternatively, this file may be used in accordance with the terms and |
| ** conditions contained in a signed written agreement between you and Nokia. |
| ** |
| ** |
| ** |
| ** |
| ** |
| ** $QT_END_LICENSE$ |
| ** |
| ****************************************************************************/ |
| |
| #ifndef QATOMIC_IA64_H |
| #define QATOMIC_IA64_H |
| |
| QT_BEGIN_HEADER |
| |
| QT_BEGIN_NAMESPACE |
| |
| #define Q_ATOMIC_INT_REFERENCE_COUNTING_IS_ALWAYS_NATIVE |
| #define Q_ATOMIC_INT_REFERENCE_COUNTING_IS_WAIT_FREE |
| |
| inline bool QBasicAtomicInt::isReferenceCountingNative() |
| { return true; } |
| inline bool QBasicAtomicInt::isReferenceCountingWaitFree() |
| { return true; } |
| |
| #define Q_ATOMIC_INT_TEST_AND_SET_IS_ALWAYS_NATIVE |
| #define Q_ATOMIC_INT_TEST_AND_SET_IS_WAIT_FREE |
| |
| inline bool QBasicAtomicInt::isTestAndSetNative() |
| { return true; } |
| inline bool QBasicAtomicInt::isTestAndSetWaitFree() |
| { return true; } |
| |
| #define Q_ATOMIC_INT_FETCH_AND_STORE_IS_ALWAYS_NATIVE |
| #define Q_ATOMIC_INT_FETCH_AND_STORE_IS_WAIT_FREE |
| |
| inline bool QBasicAtomicInt::isFetchAndStoreNative() |
| { return true; } |
| inline bool QBasicAtomicInt::isFetchAndStoreWaitFree() |
| { return true; } |
| |
| #define Q_ATOMIC_INT_FETCH_AND_ADD_IS_ALWAYS_NATIVE |
| |
| inline bool QBasicAtomicInt::isFetchAndAddNative() |
| { return true; } |
| inline bool QBasicAtomicInt::isFetchAndAddWaitFree() |
| { return false; } |
| |
| #define Q_ATOMIC_POINTER_TEST_AND_SET_IS_ALWAYS_NATIVE |
| #define Q_ATOMIC_POINTER_TEST_AND_SET_IS_WAIT_FREE |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isTestAndSetNative() |
| { return true; } |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isTestAndSetWaitFree() |
| { return true; } |
| |
| #define Q_ATOMIC_POINTER_FETCH_AND_STORE_IS_ALWAYS_NATIVE |
| #define Q_ATOMIC_POINTER_FETCH_AND_STORE_IS_WAIT_FREE |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isFetchAndStoreNative() |
| { return true; } |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isFetchAndStoreWaitFree() |
| { return true; } |
| |
| #define Q_ATOMIC_POINTER_FETCH_AND_ADD_IS_ALWAYS_NATIVE |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isFetchAndAddNative() |
| { return true; } |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::isFetchAndAddWaitFree() |
| { return false; } |
| |
| inline bool _q_ia64_fetchadd_immediate(register int value) |
| { |
| return value == 1 || value == -1 |
| || value == 4 || value == -4 |
| || value == 8 || value == -8 |
| || value == 16 || value == -16; |
| } |
| |
| #if defined(Q_CC_INTEL) |
| |
| // intrinsics provided by the Intel C++ Compiler |
| #include <ia64intrin.h> |
| |
| inline int QBasicAtomicInt::fetchAndStoreAcquire(int newValue) |
| { |
| return static_cast<int>(_InterlockedExchange(&_q_value, newValue)); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndStoreRelease(int newValue) |
| { |
| __memory_barrier(); |
| return static_cast<int>(_InterlockedExchange(&_q_value, newValue)); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetRelaxed(int expectedValue, int newValue) |
| { |
| register int expectedValueCopy = expectedValue; |
| return (static_cast<int>(_InterlockedCompareExchange(&_q_value, |
| newValue, |
| expectedValueCopy)) |
| == expectedValue); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetAcquire(int expectedValue, int newValue) |
| { |
| register int expectedValueCopy = expectedValue; |
| return (static_cast<int>(_InterlockedCompareExchange_acq(reinterpret_cast<volatile uint *>(&_q_value), |
| newValue, |
| expectedValueCopy)) |
| == expectedValue); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetRelease(int expectedValue, int newValue) |
| { |
| register int expectedValueCopy = expectedValue; |
| return (static_cast<int>(_InterlockedCompareExchange_rel(reinterpret_cast<volatile uint *>(&_q_value), |
| newValue, |
| expectedValueCopy)) |
| == expectedValue); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetOrdered(int expectedValue, int newValue) |
| { |
| __memory_barrier(); |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddAcquire(int valueToAdd) |
| { |
| if (__builtin_constant_p(valueToAdd)) { |
| if (valueToAdd == 1) |
| return __fetchadd4_acq((unsigned int *)&_q_value, 1); |
| if (valueToAdd == -1) |
| return __fetchadd4_acq((unsigned int *)&_q_value, -1); |
| } |
| return _InterlockedExchangeAdd(&_q_value, valueToAdd); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddRelease(int valueToAdd) |
| { |
| if (__builtin_constant_p(valueToAdd)) { |
| if (valueToAdd == 1) |
| return __fetchadd4_rel((unsigned int *)&_q_value, 1); |
| if (valueToAdd == -1) |
| return __fetchadd4_rel((unsigned int *)&_q_value, -1); |
| } |
| __memory_barrier(); |
| return _InterlockedExchangeAdd(&_q_value, valueToAdd); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddOrdered(int valueToAdd) |
| { |
| __memory_barrier(); |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| inline bool QBasicAtomicInt::ref() |
| { |
| return _InterlockedIncrement(&_q_value) != 0; |
| } |
| |
| inline bool QBasicAtomicInt::deref() |
| { |
| return _InterlockedDecrement(&_q_value) != 0; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreAcquire(T *newValue) |
| { |
| return (T *)_InterlockedExchangePointer(reinterpret_cast<void * volatile*>(&_q_value), newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreRelease(T *newValue) |
| { |
| __memory_barrier(); |
| return fetchAndStoreAcquire(newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetRelaxed(T *expectedValue, T *newValue) |
| { |
| register T *expectedValueCopy = expectedValue; |
| return (_InterlockedCompareExchangePointer(reinterpret_cast<void * volatile*>(&_q_value), |
| newValue, |
| expectedValueCopy) |
| == expectedValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetAcquire(T *expectedValue, T *newValue) |
| { |
| union { |
| volatile void *x; |
| volatile unsigned long *p; |
| }; |
| x = &_q_value; |
| register T *expectedValueCopy = expectedValue; |
| return (_InterlockedCompareExchange64_acq(p, quintptr(newValue), quintptr(expectedValueCopy)) |
| == quintptr(expectedValue)); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetRelease(T *expectedValue, T *newValue) |
| { |
| union { |
| volatile void *x; |
| volatile unsigned long *p; |
| }; |
| x = &_q_value; |
| register T *expectedValueCopy = expectedValue; |
| return (_InterlockedCompareExchange64_rel(p, quintptr(newValue), quintptr(expectedValueCopy)) |
| == quintptr(expectedValue)); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetOrdered(T *expectedValue, T *newValue) |
| { |
| __memory_barrier(); |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddAcquire(qptrdiff valueToAdd) |
| { |
| return (T *)_InterlockedExchangeAdd64((volatile long *)&_q_value, |
| valueToAdd * sizeof(T)); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddRelease(qptrdiff valueToAdd) |
| { |
| __memory_barrier(); |
| return (T *)_InterlockedExchangeAdd64((volatile long *)&_q_value, |
| valueToAdd * sizeof(T)); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddOrdered(qptrdiff valueToAdd) |
| { |
| __memory_barrier(); |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| #else // !Q_CC_INTEL |
| |
| # if defined(Q_CC_GNU) |
| |
| inline int QBasicAtomicInt::fetchAndStoreAcquire(int newValue) |
| { |
| int ret; |
| asm volatile("xchg4 %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (newValue) |
| : "memory"); |
| return ret; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndStoreRelease(int newValue) |
| { |
| int ret; |
| asm volatile("mf\n" |
| "xchg4 %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (newValue) |
| : "memory"); |
| return ret; |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetAcquire(int expectedValue, int newValue) |
| { |
| int ret; |
| asm volatile("mov ar.ccv=%2\n" |
| ";;\n" |
| "cmpxchg4.acq %0=%1,%3,ar.ccv\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (expectedValue), "r" (newValue) |
| : "memory"); |
| return ret == expectedValue; |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetRelease(int expectedValue, int newValue) |
| { |
| int ret; |
| asm volatile("mov ar.ccv=%2\n" |
| ";;\n" |
| "cmpxchg4.rel %0=%1,%3,ar.ccv\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (expectedValue), "r" (newValue) |
| : "memory"); |
| return ret == expectedValue; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddAcquire(int valueToAdd) |
| { |
| int ret; |
| |
| #if (__GNUC__ >= 4) |
| // We implement a fast fetch-and-add when we can |
| if (__builtin_constant_p(valueToAdd) && _q_ia64_fetchadd_immediate(valueToAdd)) { |
| asm volatile("fetchadd4.acq %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "i" (valueToAdd) |
| : "memory"); |
| return ret; |
| } |
| #endif |
| |
| // otherwise, use a loop around test-and-set |
| ret = _q_value; |
| asm volatile("0:\n" |
| " mov r9=%0\n" |
| " mov ar.ccv=%0\n" |
| " add %0=%0, %2\n" |
| " ;;\n" |
| " cmpxchg4.acq %0=%1,%0,ar.ccv\n" |
| " ;;\n" |
| " cmp.ne p6,p0 = %0, r9\n" |
| "(p6) br.dptk 0b\n" |
| "1:\n" |
| : "+r" (ret), "+m" (_q_value) |
| : "r" (valueToAdd) |
| : "r9", "p6", "memory"); |
| return ret; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddRelease(int valueToAdd) |
| { |
| int ret; |
| |
| #if (__GNUC__ >= 4) |
| // We implement a fast fetch-and-add when we can |
| if (__builtin_constant_p(valueToAdd) && _q_ia64_fetchadd_immediate(valueToAdd)) { |
| asm volatile("fetchadd4.rel %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "i" (valueToAdd) |
| : "memory"); |
| return ret; |
| } |
| #endif |
| |
| // otherwise, use a loop around test-and-set |
| ret = _q_value; |
| asm volatile("0:\n" |
| " mov r9=%0\n" |
| " mov ar.ccv=%0\n" |
| " add %0=%0, %2\n" |
| " ;;\n" |
| " cmpxchg4.rel %0=%1,%0,ar.ccv\n" |
| " ;;\n" |
| " cmp.ne p6,p0 = %0, r9\n" |
| "(p6) br.dptk 0b\n" |
| "1:\n" |
| : "+r" (ret), "+m" (_q_value) |
| : "r" (valueToAdd) |
| : "r9", "p6", "memory"); |
| return ret; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddOrdered(int valueToAdd) |
| { |
| asm volatile("mf" ::: "memory"); |
| return fetchAndAddRelease(valueToAdd); |
| } |
| |
| inline bool QBasicAtomicInt::ref() |
| { |
| int ret; |
| asm volatile("fetchadd4.acq %0=%1,1\n" |
| : "=r" (ret), "+m" (_q_value) |
| : |
| : "memory"); |
| return ret != -1; |
| } |
| |
| inline bool QBasicAtomicInt::deref() |
| { |
| int ret; |
| asm volatile("fetchadd4.rel %0=%1,-1\n" |
| : "=r" (ret), "+m" (_q_value) |
| : |
| : "memory"); |
| return ret != 1; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreAcquire(T *newValue) |
| { |
| T *ret; |
| asm volatile("xchg8 %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (newValue) |
| : "memory"); |
| return ret; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreRelease(T *newValue) |
| { |
| T *ret; |
| asm volatile("mf\n" |
| "xchg8 %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (newValue) |
| : "memory"); |
| return ret; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetAcquire(T *expectedValue, T *newValue) |
| { |
| T *ret; |
| asm volatile("mov ar.ccv=%2\n" |
| ";;\n" |
| "cmpxchg8.acq %0=%1,%3,ar.ccv\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (expectedValue), "r" (newValue) |
| : "memory"); |
| return ret == expectedValue; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetRelease(T *expectedValue, T *newValue) |
| { |
| T *ret; |
| asm volatile("mov ar.ccv=%2\n" |
| ";;\n" |
| "cmpxchg8.rel %0=%1,%3,ar.ccv\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "r" (expectedValue), "r" (newValue) |
| : "memory"); |
| return ret == expectedValue; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddAcquire(qptrdiff valueToAdd) |
| { |
| T *ret; |
| |
| #if (__GNUC__ >= 4) |
| // We implement a fast fetch-and-add when we can |
| if (__builtin_constant_p(valueToAdd) && _q_ia64_fetchadd_immediate(valueToAdd * sizeof(T))) { |
| asm volatile("fetchadd8.acq %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "i" (valueToAdd * sizeof(T)) |
| : "memory"); |
| return ret; |
| } |
| #endif |
| |
| // otherwise, use a loop around test-and-set |
| ret = _q_value; |
| asm volatile("0:\n" |
| " mov r9=%0\n" |
| " mov ar.ccv=%0\n" |
| " add %0=%0, %2\n" |
| " ;;\n" |
| " cmpxchg8.acq %0=%1,%0,ar.ccv\n" |
| " ;;\n" |
| " cmp.ne p6,p0 = %0, r9\n" |
| "(p6) br.dptk 0b\n" |
| "1:\n" |
| : "+r" (ret), "+m" (_q_value) |
| : "r" (valueToAdd * sizeof(T)) |
| : "r9", "p6", "memory"); |
| return ret; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddRelease(qptrdiff valueToAdd) |
| { |
| T *ret; |
| |
| #if (__GNUC__ >= 4) |
| // We implement a fast fetch-and-add when we can |
| if (__builtin_constant_p(valueToAdd) && _q_ia64_fetchadd_immediate(valueToAdd * sizeof(T))) { |
| asm volatile("fetchadd8.rel %0=%1,%2\n" |
| : "=r" (ret), "+m" (_q_value) |
| : "i" (valueToAdd * sizeof(T)) |
| : "memory"); |
| return ret; |
| } |
| #endif |
| |
| // otherwise, use a loop around test-and-set |
| ret = _q_value; |
| asm volatile("0:\n" |
| " mov r9=%0\n" |
| " mov ar.ccv=%0\n" |
| " add %0=%0, %2\n" |
| " ;;\n" |
| " cmpxchg8.rel %0=%1,%0,ar.ccv\n" |
| " ;;\n" |
| " cmp.ne p6,p0 = %0, r9\n" |
| "(p6) br.dptk 0b\n" |
| "1:\n" |
| : "+r" (ret), "+m" (_q_value) |
| : "r" (valueToAdd * sizeof(T)) |
| : "r9", "p6", "memory"); |
| return ret; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddOrdered(qptrdiff valueToAdd) |
| { |
| asm volatile("mf" ::: "memory"); |
| return fetchAndAddRelease(valueToAdd); |
| } |
| |
| #elif defined Q_CC_HPACC |
| |
| QT_BEGIN_INCLUDE_NAMESPACE |
| #include <ia64/sys/inline.h> |
| QT_END_INCLUDE_NAMESPACE |
| |
| #define FENCE (_Asm_fence)(_UP_CALL_FENCE | _UP_SYS_FENCE | _DOWN_CALL_FENCE | _DOWN_SYS_FENCE) |
| |
| inline int QBasicAtomicInt::fetchAndStoreAcquire(int newValue) |
| { |
| return _Asm_xchg((_Asm_sz)_SZ_W, &_q_value, (unsigned)newValue, |
| (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndStoreRelease(int newValue) |
| { |
| _Asm_mf(FENCE); |
| return _Asm_xchg((_Asm_sz)_SZ_W, &_q_value, (unsigned)newValue, |
| (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetAcquire(int expectedValue, int newValue) |
| { |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (unsigned)expectedValue, FENCE); |
| int ret = _Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, (unsigned)newValue, (_Asm_ldhint)_LDHINT_NONE); |
| return ret == expectedValue; |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetRelease(int expectedValue, int newValue) |
| { |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (unsigned)expectedValue, FENCE); |
| int ret = _Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, newValue, (_Asm_ldhint)_LDHINT_NONE); |
| return ret == expectedValue; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddAcquire(int valueToAdd) |
| { |
| if (valueToAdd == 1) |
| return _Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, 1, (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| else if (valueToAdd == -1) |
| return _Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, -1, (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| |
| // implement the test-and-set loop |
| register int old, ret; |
| do { |
| old = _q_value; |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (unsigned)old, FENCE); |
| ret = _Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, old + valueToAdd, (_Asm_ldhint)_LDHINT_NONE); |
| } while (ret != old); |
| return old; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddRelease(int valueToAdd) |
| { |
| if (valueToAdd == 1) |
| return _Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, 1, (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| else if (valueToAdd == -1) |
| return _Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, -1, (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| |
| // implement the test-and-set loop |
| register int old, ret; |
| do { |
| old = _q_value; |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (unsigned)old, FENCE); |
| ret = _Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, old + valueToAdd, (_Asm_ldhint)_LDHINT_NONE); |
| } while (ret != old); |
| return old; |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddOrdered(int valueToAdd) |
| { |
| _Asm_mf(FENCE); |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| inline bool QBasicAtomicInt::ref() |
| { |
| return (int)_Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, 1, (_Asm_ldhint)_LDHINT_NONE, FENCE) != -1; |
| } |
| |
| inline bool QBasicAtomicInt::deref() |
| { |
| return (int)_Asm_fetchadd((_Asm_fasz)_FASZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, -1, (_Asm_ldhint)_LDHINT_NONE, FENCE) != 1; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreAcquire(T *newValue) |
| { |
| #ifdef __LP64__ |
| return (T *)_Asm_xchg((_Asm_sz)_SZ_D, &_q_value, (quint64)newValue, |
| (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| #else |
| return (T *)_Asm_xchg((_Asm_sz)_SZ_W, &_q_value, (quint32)newValue, |
| (_Asm_ldhint)_LDHINT_NONE, FENCE); |
| #endif |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreRelease(T *newValue) |
| { |
| _Asm_mf(FENCE); |
| return fetchAndStoreAcquire(newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetAcquire(T *expectedValue, T *newValue) |
| { |
| #ifdef __LP64__ |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint64)expectedValue, FENCE); |
| T *ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_D, (_Asm_sem)_SEM_ACQ, |
| &_q_value, (quint64)newValue, (_Asm_ldhint)_LDHINT_NONE); |
| #else |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint32)expectedValue, FENCE); |
| T *ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, (quint32)newValue, (_Asm_ldhint)_LDHINT_NONE); |
| #endif |
| return ret == expectedValue; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetRelease(T *expectedValue, T *newValue) |
| { |
| #ifdef __LP64__ |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint64)expectedValue, FENCE); |
| T *ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_D, (_Asm_sem)_SEM_REL, |
| &_q_value, (quint64)newValue, (_Asm_ldhint)_LDHINT_NONE); |
| #else |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint32)expectedValue, FENCE); |
| T *ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, (quint32)newValue, (_Asm_ldhint)_LDHINT_NONE); |
| #endif |
| return ret == expectedValue; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddAcquire(qptrdiff valueToAdd) |
| { |
| // implement the test-and-set loop |
| register T *old, *ret; |
| do { |
| old = _q_value; |
| #ifdef __LP64__ |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint64)old, FENCE); |
| ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_D, (_Asm_sem)_SEM_ACQ, |
| &_q_value, (quint64)(old + valueToAdd), |
| (_Asm_ldhint)_LDHINT_NONE); |
| #else |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint32)old, FENCE); |
| ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_ACQ, |
| &_q_value, (quint32)(old + valueToAdd), |
| (_Asm_ldhint)_LDHINT_NONE); |
| #endif |
| } while (old != ret); |
| return old; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddRelease(qptrdiff valueToAdd) |
| { |
| // implement the test-and-set loop |
| register T *old, *ret; |
| do { |
| old = _q_value; |
| #ifdef __LP64__ |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint64)old, FENCE); |
| ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_D, (_Asm_sem)_SEM_REL, |
| &_q_value, (quint64)(old + valueToAdd), |
| (_Asm_ldhint)_LDHINT_NONE); |
| #else |
| _Asm_mov_to_ar((_Asm_app_reg)_AREG_CCV, (quint32)old, FENCE); |
| ret = (T *)_Asm_cmpxchg((_Asm_sz)_SZ_W, (_Asm_sem)_SEM_REL, |
| &_q_value, (quint32)(old + valueToAdd), |
| (_Asm_ldhint)_LDHINT_NONE); |
| #endif |
| } while (old != ret); |
| return old; |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddOrdered(qptrdiff valueToAdd) |
| { |
| _Asm_mf(FENCE); |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| #else |
| |
| extern "C" { |
| Q_CORE_EXPORT int q_atomic_test_and_set_int(volatile int *ptr, int expected, int newval); |
| Q_CORE_EXPORT int q_atomic_test_and_set_ptr(volatile void *ptr, void *expected, void *newval); |
| } // extern "C" |
| |
| #endif |
| |
| inline bool QBasicAtomicInt::testAndSetRelaxed(int expectedValue, int newValue) |
| { |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| inline bool QBasicAtomicInt::testAndSetOrdered(int expectedValue, int newValue) |
| { |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetRelaxed(T *expectedValue, T *newValue) |
| { |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE bool QBasicAtomicPointer<T>::testAndSetOrdered(T *expectedValue, T *newValue) |
| { |
| return testAndSetAcquire(expectedValue, newValue); |
| } |
| |
| #endif // Q_CC_INTEL |
| |
| inline int QBasicAtomicInt::fetchAndStoreRelaxed(int newValue) |
| { |
| return fetchAndStoreAcquire(newValue); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndStoreOrdered(int newValue) |
| { |
| return fetchAndStoreRelease(newValue); |
| } |
| |
| inline int QBasicAtomicInt::fetchAndAddRelaxed(int valueToAdd) |
| { |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreRelaxed(T *newValue) |
| { |
| return fetchAndStoreAcquire(newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndStoreOrdered(T *newValue) |
| { |
| return fetchAndStoreRelaxed(newValue); |
| } |
| |
| template <typename T> |
| Q_INLINE_TEMPLATE T *QBasicAtomicPointer<T>::fetchAndAddRelaxed(qptrdiff valueToAdd) |
| { |
| return fetchAndAddAcquire(valueToAdd); |
| } |
| |
| QT_END_NAMESPACE |
| |
| QT_END_HEADER |
| |
| #endif // QATOMIC_IA64_H |