|| ||Ingo Molnar <mingo-AT-elte.hu>|
|| ||Linus Torvalds <torvalds-AT-osdl.org>|
|| ||Re: [patch] remove the BKL (Big Kernel Lock), this time for real|
|| ||Wed, 15 Sep 2004 17:55:55 +0200|
|| ||linux-kernel-AT-vger.kernel.org, Andrew Morton <akpm-AT-osdl.org>,
William Lee Irwin III <wli-AT-holomorphy.com>,
Arjan van de Ven <arjanv-AT-redhat.com>, Lee Revell <rlrevell-AT-joe-job.com>|
* Linus Torvalds <email@example.com> wrote:
> > the attached patch is a new approach to get rid of Linux's Big Kernel
> > Lock as we know it today.
> I really think this is wrong.
> Maybe not from a conceptual standpoint, but that implementation with
> the scheduler doing "reaquire_kernel_lock()" and doing a down() there
> is just wrong, wrong, wrong.
> If we're going to do a down() and block immediately after being
> scheduled, I don't think we should have been picked in the first
agreed, and that codepath is only for correctness - the semaphore code
will make sure most of the time that this doesnt happen. It's the same
as the need_resched check - it doesnt happen in 99.9% of the cases but
when it happens it must happen.
> Yeah, yeah, you have all that magic to not recurse by setting
> lock-depth negative before doing the down(), but it still feels
> fundamentally wrong to me. There's also the question whether this
> actually _helps_ anything, since it may well just replace the spinning
> with lots of new scheduler activity.
Rare activity that still runs under the BKL (e.g. mounting, or ioctls)
can introduce many milliseconds of scheduling delays that hurt
latencies. None of this is really performance-sensitive as it's almost
always used for some big old piece of code. Anything that is frequently
taken/dropped we've replaced with proper spinlocks already. So what's
left is the code for which 1) everyone hurts most from it not being a
real semaphore 2) no _user_ or maintainer of that code cares about the
BKL because it's not performance-critical.
> And you make schedule() a lot more expensive for kernel lock holders
> by copying the CPU map. [...]
we dont really care about BKL _users_, other than correctness. We've got
cpusets for the bitmap overhead reduction.
> [...] You may have tested it on a machine where the CPU map is just a
> single word, but what about the big machines?
and it's not that bad - if it's .. 512 CPUs then _BKL users_ copy 64
bytes. I dont think the 512-CPU guys will complain about this patch. If
there are any frequent BKL users on such big boxes then they need
serious and quick fixing anyway, because they are bouncing a global
spinlock madly across 512 CPUs, 32 cross-connects and 4 backplanes! The
64-byte copy within the CPU-local task structure really dwarves in
> Spinlocks really _are_ cheaper. Wouldn't it be nice to just continue
> removing kernel lock users and keeping a very _simple_ kernel lock for
> legacy issues?
yes, but progress in this area seems to have slowed down, and people are
hurting from the latencies introduced by the BKL meanwhile. Who cares if
some rare big chunk of code runs under a semaphore, as long as it's
a global spinlock isnt all that much cheaper than a semaphore, even if
it's not taken or lightly taken. If it's heavily taken then a semaphore
wins. There is clearly an interim area where a spinlock will win - but
only if the spinlock is well-localized to some resource or CPU - this is
absolutely not true for the BKL which is as global as it gets.
> In other words, I'd _really_ like to see some serious numbers for
fully agreed about that too!
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to firstname.lastname@example.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
to post comments)