On Fri, 27 May 2005 21:04:37 -0400, Kyle Moffett <[email protected]>
wrote:
On May 27, 2005, at 18:31:38, David Nicol wrote:
On 5/26/05, john cooper <[email protected]> wrote:
given design. Clearly we aren't buying anything to trade off
a spinlock protecting the update of a single pointer with a
blocking lock and associated context switching.
On contention, and only on contention, we are faced with the question
of what
to do. Do we wait, or do we go away and come back later? What
information is
available to us to base the decision on? We can't gather any more
information,
because that would take longer than spin-waiting. If the "spinaphore"
told us,
on acquisition failure, how many other threads were asking for it, we
could implement
a tunable lock, that surrenders context when there are more than N
threads waiting for
the resource, and that otherwise waits its turn, or its chance, as a
compromise
and synthesis.
Here is an example naive implementation which could perhaps be optimized
further
for architectures based on memory and synchronization requirements.
A quick summary:
Each time the lock is taken and released, a "hold_time" is updated which
indicates
the average time that the lock is held. During contention, each CPU
checks the
current average hold time and the number of CPUs waiting against a
predefined
"context switch + useful work" time, and goes to sleep if it thinks it
has enough
time to spare.
If you went with a bakery algorithm and could tolerate FIFO service order,
you could use
the expected service time as the ticket increment value instead of 1.
Before a thread
gets a ticket, it examines the expected queue wait time, the difference
between the
current ticket and the next available ticket, to decide which increment to
be applied
to the next ticket value. The two possible increment values would be the
uncontended
resource service time and that value plus thread suspend/resume
overhead. If the
expected wait time is greater than the latter, it uses the latter as the
increment value
and suspends rather than spins.
Bakery algorithms have other nice properties. The lock release
(incrementing the
current ticket) doesn't require an interlocked operation, though the
release memory
barrier and other memory barriers required to determine if there are any
waiters may
make that somewhat moot. The next and current tickets could be kept in
separate
cache lines. And the get ticket interlocked operations are staggered
out, unlike a
conventional spin lock where once the lock is released, *all* waiting
processors
attempt to acquire the lock cache line exclusive all at once. The slows
down
the lock release since the cache line is guaranteed to be held by another
processor
if there was lock contention.
Also bakery spin locks can be rather easily modified to be rwlocks with no
extra
overhead to speak of. Basically, you get rwlock capability for free.
--
Joe Seigh
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]