mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-30 13:38:40 +00:00
3a6bfbc91d
The arch_mutex_cpu_relax() function, introduced by 34b133f
, is
hacky and ugly. It was added a few years ago to address the fact
that common cpu_relax() calls include yielding on s390, and thus
impact the optimistic spinning functionality of mutexes. Nowadays
we use this function well beyond mutexes: rwsem, qrwlock, mcs and
lockref. Since the macro that defines the call is in the mutex header,
any users must include mutex.h and the naming is misleading as well.
This patch (i) renames the call to cpu_relax_lowlatency ("relax, but
only if you can do it with very low latency") and (ii) defines it in
each arch's asm/processor.h local header, just like for regular cpu_relax
functions. On all archs, except s390, cpu_relax_lowlatency is simply cpu_relax,
and thus we can take it out of mutex.h. While this can seem redundant,
I believe it is a good choice as it allows us to move out arch specific
logic from generic locking primitives and enables future(?) archs to
transparently define it, similarly to System Z.
Signed-off-by: Davidlohr Bueso <davidlohr@hp.com>
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Anton Blanchard <anton@samba.org>
Cc: Aurelien Jacquiot <a-jacquiot@ti.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Bharat Bhushan <r65777@freescale.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Chen Liqin <liqin.linux@gmail.com>
Cc: Chris Metcalf <cmetcalf@tilera.com>
Cc: Christian Borntraeger <borntraeger@de.ibm.com>
Cc: Chris Zankel <chris@zankel.net>
Cc: David Howells <dhowells@redhat.com>
Cc: David S. Miller <davem@davemloft.net>
Cc: Deepthi Dharwar <deepthi@linux.vnet.ibm.com>
Cc: Dominik Dingel <dingel@linux.vnet.ibm.com>
Cc: Fenghua Yu <fenghua.yu@intel.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Guan Xuetao <gxt@mprc.pku.edu.cn>
Cc: Haavard Skinnemoen <hskinnemoen@gmail.com>
Cc: Hans-Christian Egtvedt <egtvedt@samfundet.no>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Helge Deller <deller@gmx.de>
Cc: Hirokazu Takata <takata@linux-m32r.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: James E.J. Bottomley <jejb@parisc-linux.org>
Cc: James Hogan <james.hogan@imgtec.com>
Cc: Jason Wang <jasowang@redhat.com>
Cc: Jesper Nilsson <jesper.nilsson@axis.com>
Cc: Joe Perches <joe@perches.com>
Cc: Jonas Bonn <jonas@southpole.se>
Cc: Joseph Myers <joseph@codesourcery.com>
Cc: Kees Cook <keescook@chromium.org>
Cc: Koichi Yasutake <yasutake.koichi@jp.panasonic.com>
Cc: Lennox Wu <lennox.wu@gmail.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mark Salter <msalter@redhat.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Max Filippov <jcmvbkbc@gmail.com>
Cc: Michael Neuling <mikey@neuling.org>
Cc: Michal Simek <monstr@monstr.eu>
Cc: Mikael Starvik <starvik@axis.com>
Cc: Nicolas Pitre <nico@linaro.org>
Cc: Paolo Bonzini <pbonzini@redhat.com>
Cc: Paul Burton <paul.burton@imgtec.com>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Paul Gortmaker <paul.gortmaker@windriver.com>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Qais Yousef <qais.yousef@imgtec.com>
Cc: Qiaowei Ren <qiaowei.ren@intel.com>
Cc: Rafael Wysocki <rafael.j.wysocki@intel.com>
Cc: Ralf Baechle <ralf@linux-mips.org>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Richard Kuo <rkuo@codeaurora.org>
Cc: Russell King <linux@arm.linux.org.uk>
Cc: Steven Miao <realmz6@gmail.com>
Cc: Steven Rostedt <srostedt@redhat.com>
Cc: Stratos Karafotis <stratosk@semaphore.gr>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Cc: Tony Luck <tony.luck@intel.com>
Cc: Vasily Kulikov <segoon@openwall.com>
Cc: Vineet Gupta <vgupta@synopsys.com>
Cc: Vineet Gupta <Vineet.Gupta1@synopsys.com>
Cc: Waiman Long <Waiman.Long@hp.com>
Cc: Will Deacon <will.deacon@arm.com>
Cc: Wolfram Sang <wsa@the-dreams.de>
Cc: adi-buildroot-devel@lists.sourceforge.net
Cc: linux390@de.ibm.com
Cc: linux-alpha@vger.kernel.org
Cc: linux-am33-list@redhat.com
Cc: linux-arm-kernel@lists.infradead.org
Cc: linux-c6x-dev@linux-c6x.org
Cc: linux-cris-kernel@axis.com
Cc: linux-hexagon@vger.kernel.org
Cc: linux-ia64@vger.kernel.org
Cc: linux@lists.openrisc.net
Cc: linux-m32r-ja@ml.linux-m32r.org
Cc: linux-m32r@ml.linux-m32r.org
Cc: linux-m68k@lists.linux-m68k.org
Cc: linux-metag@vger.kernel.org
Cc: linux-mips@linux-mips.org
Cc: linux-parisc@vger.kernel.org
Cc: linuxppc-dev@lists.ozlabs.org
Cc: linux-s390@vger.kernel.org
Cc: linux-sh@vger.kernel.org
Cc: linux-xtensa@linux-xtensa.org
Cc: sparclinux@vger.kernel.org
Link: http://lkml.kernel.org/r/1404079773.2619.4.camel@buesod1.americas.hpqcorp.net
Signed-off-by: Ingo Molnar <mingo@kernel.org>
133 lines
3.6 KiB
C
133 lines
3.6 KiB
C
/*
|
|
* Queue read/write lock
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* (C) Copyright 2013-2014 Hewlett-Packard Development Company, L.P.
|
|
*
|
|
* Authors: Waiman Long <waiman.long@hp.com>
|
|
*/
|
|
#include <linux/smp.h>
|
|
#include <linux/bug.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/hardirq.h>
|
|
#include <asm/qrwlock.h>
|
|
|
|
/**
|
|
* rspin_until_writer_unlock - inc reader count & spin until writer is gone
|
|
* @lock : Pointer to queue rwlock structure
|
|
* @writer: Current queue rwlock writer status byte
|
|
*
|
|
* In interrupt context or at the head of the queue, the reader will just
|
|
* increment the reader count & wait until the writer releases the lock.
|
|
*/
|
|
static __always_inline void
|
|
rspin_until_writer_unlock(struct qrwlock *lock, u32 cnts)
|
|
{
|
|
while ((cnts & _QW_WMASK) == _QW_LOCKED) {
|
|
cpu_relax_lowlatency();
|
|
cnts = smp_load_acquire((u32 *)&lock->cnts);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* queue_read_lock_slowpath - acquire read lock of a queue rwlock
|
|
* @lock: Pointer to queue rwlock structure
|
|
*/
|
|
void queue_read_lock_slowpath(struct qrwlock *lock)
|
|
{
|
|
u32 cnts;
|
|
|
|
/*
|
|
* Readers come here when they cannot get the lock without waiting
|
|
*/
|
|
if (unlikely(in_interrupt())) {
|
|
/*
|
|
* Readers in interrupt context will spin until the lock is
|
|
* available without waiting in the queue.
|
|
*/
|
|
cnts = smp_load_acquire((u32 *)&lock->cnts);
|
|
rspin_until_writer_unlock(lock, cnts);
|
|
return;
|
|
}
|
|
atomic_sub(_QR_BIAS, &lock->cnts);
|
|
|
|
/*
|
|
* Put the reader into the wait queue
|
|
*/
|
|
arch_spin_lock(&lock->lock);
|
|
|
|
/*
|
|
* At the head of the wait queue now, wait until the writer state
|
|
* goes to 0 and then try to increment the reader count and get
|
|
* the lock. It is possible that an incoming writer may steal the
|
|
* lock in the interim, so it is necessary to check the writer byte
|
|
* to make sure that the write lock isn't taken.
|
|
*/
|
|
while (atomic_read(&lock->cnts) & _QW_WMASK)
|
|
cpu_relax_lowlatency();
|
|
|
|
cnts = atomic_add_return(_QR_BIAS, &lock->cnts) - _QR_BIAS;
|
|
rspin_until_writer_unlock(lock, cnts);
|
|
|
|
/*
|
|
* Signal the next one in queue to become queue head
|
|
*/
|
|
arch_spin_unlock(&lock->lock);
|
|
}
|
|
EXPORT_SYMBOL(queue_read_lock_slowpath);
|
|
|
|
/**
|
|
* queue_write_lock_slowpath - acquire write lock of a queue rwlock
|
|
* @lock : Pointer to queue rwlock structure
|
|
*/
|
|
void queue_write_lock_slowpath(struct qrwlock *lock)
|
|
{
|
|
u32 cnts;
|
|
|
|
/* Put the writer into the wait queue */
|
|
arch_spin_lock(&lock->lock);
|
|
|
|
/* Try to acquire the lock directly if no reader is present */
|
|
if (!atomic_read(&lock->cnts) &&
|
|
(atomic_cmpxchg(&lock->cnts, 0, _QW_LOCKED) == 0))
|
|
goto unlock;
|
|
|
|
/*
|
|
* Set the waiting flag to notify readers that a writer is pending,
|
|
* or wait for a previous writer to go away.
|
|
*/
|
|
for (;;) {
|
|
cnts = atomic_read(&lock->cnts);
|
|
if (!(cnts & _QW_WMASK) &&
|
|
(atomic_cmpxchg(&lock->cnts, cnts,
|
|
cnts | _QW_WAITING) == cnts))
|
|
break;
|
|
|
|
cpu_relax_lowlatency();
|
|
}
|
|
|
|
/* When no more readers, set the locked flag */
|
|
for (;;) {
|
|
cnts = atomic_read(&lock->cnts);
|
|
if ((cnts == _QW_WAITING) &&
|
|
(atomic_cmpxchg(&lock->cnts, _QW_WAITING,
|
|
_QW_LOCKED) == _QW_WAITING))
|
|
break;
|
|
|
|
cpu_relax_lowlatency();
|
|
}
|
|
unlock:
|
|
arch_spin_unlock(&lock->lock);
|
|
}
|
|
EXPORT_SYMBOL(queue_write_lock_slowpath);
|