summaryrefslogtreecommitdiffstats
path: root/fs/bio.c
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2010-10-27 12:38:12 -0400
committerArnd Bergmann <arnd@arndb.de>2010-10-27 22:06:17 +0200
commitf7347ce4ee7c65415f84be915c018473e7076f31 (patch)
tree613ce14f088ad00bdbc77cdfb686a40a4851180f /fs/bio.c
parentc5b1f0d92c36851aca09ac6c7c0c4f9690ac14f3 (diff)
downloadlinux-f7347ce4ee7c65415f84be915c018473e7076f31.tar.gz
linux-f7347ce4ee7c65415f84be915c018473e7076f31.tar.bz2
linux-f7347ce4ee7c65415f84be915c018473e7076f31.zip
fasync: re-organize fasync entry insertion to allow it under a spinlock
You currently cannot use "fasync_helper()" in an atomic environment to insert a new fasync entry, because it will need to allocate the new "struct fasync_struct". Yet fcntl_setlease() wants to call this under lock_flocks(), which is in the process of being converted from the BKL to a spinlock. In order to fix this, this abstracts out the actual fasync list insertion and the fasync allocations into functions of their own, and teaches fs/locks.c to pre-allocate the fasync_struct entry. That way the actual list insertion can happen while holding the required spinlock. Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org> [bfields@redhat.com: rebase on top of my changes to Arnd's patch] Tested-by: J. Bruce Fields <bfields@redhat.com> Signed-off-by: Arnd Bergmann <arnd@arndb.de>
Diffstat (limited to 'fs/bio.c')
0 files changed, 0 insertions, 0 deletions