diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2015-09-17 08:45:23 -0700 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2015-09-17 08:45:23 -0700 |
commit | 9786cff38a31c452e32fd2f9a479dd7d19d91712 (patch) | |
tree | 1197c2e7a4e8973d19919f54f342fba6067488e2 /arch/x86/include | |
parent | 1b3dfde386b7c72b8f5430dc40eee538eb40c948 (diff) | |
parent | 1975dbc276c6ab62230cf4f9df5ddc9ff0e0e473 (diff) | |
download | linux-9786cff38a31c452e32fd2f9a479dd7d19d91712.tar.gz linux-9786cff38a31c452e32fd2f9a479dd7d19d91712.tar.bz2 linux-9786cff38a31c452e32fd2f9a479dd7d19d91712.zip |
Merge branch 'locking-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull locking fixes from Ingo Molnar:
"Spinlock performance regression fix, plus documentation fixes"
* 'locking-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
locking/static_keys: Fix up the static keys documentation
locking/qspinlock/x86: Only emit the test-and-set fallback when building guest support
locking/qspinlock/x86: Fix performance regression under unaccelerated VMs
locking/static_keys: Fix a silly typo
Diffstat (limited to 'arch/x86/include')
-rw-r--r-- | arch/x86/include/asm/qspinlock.h | 19 |
1 files changed, 14 insertions, 5 deletions
diff --git a/arch/x86/include/asm/qspinlock.h b/arch/x86/include/asm/qspinlock.h index 9d51fae1cba3..eaba08076030 100644 --- a/arch/x86/include/asm/qspinlock.h +++ b/arch/x86/include/asm/qspinlock.h @@ -39,18 +39,27 @@ static inline void queued_spin_unlock(struct qspinlock *lock) } #endif -#define virt_queued_spin_lock virt_queued_spin_lock - -static inline bool virt_queued_spin_lock(struct qspinlock *lock) +#ifdef CONFIG_PARAVIRT +#define virt_spin_lock virt_spin_lock +static inline bool virt_spin_lock(struct qspinlock *lock) { if (!static_cpu_has(X86_FEATURE_HYPERVISOR)) return false; - while (atomic_cmpxchg(&lock->val, 0, _Q_LOCKED_VAL) != 0) - cpu_relax(); + /* + * On hypervisors without PARAVIRT_SPINLOCKS support we fall + * back to a Test-and-Set spinlock, because fair locks have + * horrible lock 'holder' preemption issues. + */ + + do { + while (atomic_read(&lock->val) != 0) + cpu_relax(); + } while (atomic_cmpxchg(&lock->val, 0, _Q_LOCKED_VAL) != 0); return true; } +#endif /* CONFIG_PARAVIRT */ #include <asm-generic/qspinlock.h> |