1. 28 3月, 2006 40 次提交
    • J
      [PATCH] [BLOCK] cfq-iosched: change cfq io context linking from list to tree · e2d74ac0
      Jens Axboe 提交于
      On setups with many disks, we spend a considerable amount of time
      looking up the process-disk mapping on each queue of io. Testing with
      a NULL based block driver, this costs 40-50% reduction in throughput
      for 1000 disks.
      Signed-off-by: NJens Axboe <axboe@suse.de>
      e2d74ac0
    • N
      [PATCH] md: Convert reconfig_sem to reconfig_mutex · df5b89b3
      NeilBrown 提交于
      ... being careful that mutex_trylock is inverted wrt down_trylock
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      df5b89b3
    • N
      [PATCH] md: Support suspending of IO to regions of an md array · e464eafd
      NeilBrown 提交于
      This allows user-space to access data safely.  This is needed for raid5
      reshape as user-space needs to take a backup of the first few stripes before
      allowing reshape to commence.
      
      It will also be useful in cluster-aware raid1 configurations so that all
      cluster members can leave a section of the array untouched while a
      resync/recovery happens.
      
      A 'start' and 'end' of the suspended range are written to 2 sysfs attributes.
      Note that only one range can be suspended at a time.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      e464eafd
    • N
      [PATCH] md: Split reshape handler in check_reshape and start_reshape · 63c70c4f
      NeilBrown 提交于
      check_reshape checks validity and does things that can be done instantly -
      like adding devices to raid1.  start_reshape initiates a restriping process to
      convert the whole array.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      63c70c4f
    • N
      [PATCH] md: Only checkpoint expansion progress occasionally · b578d55f
      NeilBrown 提交于
      Instead of checkpointing at each stripe, only checkpoint when a new write
      would overwrite uncheckpointed data.  Block any write to the uncheckpointed
      area.  Arbitrarily checkpoint at least every 3Meg.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      b578d55f
    • N
      [PATCH] md: Checkpoint and allow restart of raid5 reshape · f6705578
      NeilBrown 提交于
      We allow the superblock to record an 'old' and a 'new' geometry, and a
      position where any conversion is up to.  The geometry allows for changing
      chunksize, layout and level as well as number of devices.
      
      When using verion-0.90 superblock, we convert the version to 0.91 while the
      conversion is happening so that an old kernel will refuse the assemble the
      array.  For version-1, we use a feature bit for the same effect.
      
      When starting an array we check for an incomplete reshape and restart the
      reshape process if needed.  If the reshape stopped at an awkward time (like
      when updating the first stripe) we refuse to assemble the array, and let
      user-space worry about it.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      f6705578
    • N
      [PATCH] md: Final stages of raid5 expand code · 29269553
      NeilBrown 提交于
      This patch adds raid5_reshape and end_reshape which will start and finish the
      reshape processes.
      
      raid5_reshape is only enabled in CONFIG_MD_RAID5_RESHAPE is set, to discourage
      accidental use.
      
      Read the 'help' for the CONFIG_MD_RAID5_RESHAPE entry.
      
      and Make sure that you have backups, just in case.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      29269553
    • N
      [PATCH] md: Core of raid5 resize process · ccfcc3c1
      NeilBrown 提交于
      This patch provides the core of the resize/expand process.
      
      sync_request notices if a 'reshape' is happening and acts accordingly.
      
      It allocated new stripe_heads for the next chunk-wide-stripe in the target
      geometry, marking them STRIPE_EXPANDING.
      
      Then it finds which stripe heads in the old geometry can provide data needed
      by these and marks them STRIPE_EXPAND_SOURCE.  This causes stripe_handle to
      read all blocks on those stripes.
      
      Once all blocks on a STRIPE_EXPAND_SOURCE stripe_head are read, any that are
      needed are copied into the corresponding STRIPE_EXPANDING stripe_head.  Once a
      STRIPE_EXPANDING stripe_head is full, it is marks STRIPE_EXPAND_READY and then
      is written out and released.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      ccfcc3c1
    • N
      [PATCH] md: Infrastructure to allow normal IO to continue while array is expanding · 7ecaa1e6
      NeilBrown 提交于
      We need to allow that different stripes are of different effective sizes, and
      use the appropriate size.  Also, when a stripe is being expanded, we must
      block any IO attempts until the stripe is stable again.
      
      Key elements in this change are:
       - each stripe_head gets a 'disk' field which is part of the key,
         thus there can sometimes be two stripe heads of the same area of
         the array, but covering different numbers of devices.  One of these
         will be marked STRIPE_EXPANDING and so won't accept new requests.
       - conf->expand_progress tracks how the expansion is progressing and
         is used to determine whether the target part of the array has been
         expanded yet or not.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      7ecaa1e6
    • N
      [PATCH] md: Allow stripes to be expanded in preparation for expanding an array · ad01c9e3
      NeilBrown 提交于
      Before a RAID-5 can be expanded, we need to be able to expand the stripe-cache
      data structure.
      
      This requires allocating new stripes in a new kmem_cache.  If this succeeds,
      we copy cache pages over and release the old stripes and kmem_cache.
      
      We then allocate new pages.  If that fails, we leave the stripe cache at it's
      new size.  It isn't worth the effort to shrink it back again.
      
      Unfortuanately this means we need two kmem_cache names as we, for a short
      period of time, we have two kmem_caches.  So they are raid5/%s and
      raid5/%s-alt
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      ad01c9e3
    • N
      [PATCH] md: Split disks array out of raid5 conf structure so it is easier to grow · b55e6bfc
      NeilBrown 提交于
      The remainder of this batch implements raid5 reshaping.  Currently the only
      shape change that is supported is added a device, but it is envisioned that
      changing the chunksize and layout will also be supported, as well as changing
      the level (e.g.  1->5, 5->6).
      
      The reshape process naturally has to move all of the data in the array, and so
      should be used with caution.  It is believed to work, and some testing does
      support this, but wider testing would be great for increasing my confidence.
      
      You will need a version of mdadm newer than 2.3.1 to make use of raid5 growth.
       This is because mdadm need to take a copy of a 'critical section' at the
      start of the array incase there is a crash at an awkward moment.  On restart,
      mdadm will restore the critical section and allow reshape to continue.
      
      I hope to release a 2.4-pre by early next week - it still needs a little more
      polishing.
      
      This patch:
      
      Previously the array of disk information was included in the raid5 'conf'
      structure which was allocated to an appropriate size.  This makes it awkward
      to change the size of that array.  So we split it off into a separate
      kmalloced array which will require a little extra indexing, but is much easier
      to grow.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      b55e6bfc
    • J
      [PATCH] dm/md dependency tree in sysfs: bd_claim_by_kobject · 641dc636
      Jun'ichi Nomura 提交于
      Adding bd_claim_by_kobject() function which takes kobject as additional
      signature of holder device and creates sysfs symlinks between holder device
      and claimed device.  bd_release_from_kobject() is a counterpart of
      bd_claim_by_kobject.
      Signed-off-by: NJun'ichi Nomura <j-nomura@ce.jp.nec.com>
      Cc: Alasdair G Kergon <agk@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      641dc636
    • A
      [PATCH] dm-md-dependency-tree-in-sysfs-holders-slaves-subdirectory-tidy · 10087368
      Andrew Morton 提交于
      Remove all the CONFIG_SYSFS stuff.  That's supposed to all be implemented up
      in header files.
      
      Yes, the CONFIG_SYSFS=n data structures will be a little larger than
      necessary, but that's a tradeoff we can decide to make.
      
      Cc: Jun'ichi Nomura <j-nomura@ce.jp.nec.com>
      Cc: Alasdair G Kergon <agk@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      10087368
    • J
      [PATCH] dm/md dependency tree in sysfs: holders/slaves subdirectory · 6a4d44c1
      Jun'ichi Nomura 提交于
      Creating "slaves" and "holders" directories in /sys/block/<disk> and
      creating "holders" directory under /sys/block/<disk>/<partition>
      Signed-off-by: NJun'ichi Nomura <j-nomura@ce.jp.nec.com>
      Cc: Alasdair G Kergon <agk@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      6a4d44c1
    • D
      [PATCH] dm store geometry · 3ac51e74
      Darrick J. Wong 提交于
      Allow drive geometry to be stored with a new DM_DEV_SET_GEOMETRY ioctl.
      Device-mapper will now respond to HDIO_GETGEO.  If the geometry information is
      not available, zero will be returned for all of the parameters.
      Signed-off-by: NDarrick J. Wong <djwong@us.ibm.com>
      Signed-off-by: NAlasdair G Kergon <agk@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      3ac51e74
    • N
      [PATCH] dm: make sure QUEUE_FLAG_CLUSTER is set properly · 969429b5
      NeilBrown 提交于
      This flag should be set for a virtual device iff it is set for all
      underlying devices.
      Signed-off-by: NNeil Brown <neilb@suse.de>
      Acked-by: NAlasdair G Kergon <agk@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      969429b5
    • P
      [PATCH] Add ID for Quadro NVS280 · ed49843b
      Pavel Roskin 提交于
      Quadro NVS280 is a dual-head PCIe card with PCI ID 10de:00fd and subsystem ID
      10de:0215.
      Signed-off-by: NPavel Roskin <proski@gnu.org>
      Signed-off-by: NAntonino Daplas <adaplas@pol.net>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      ed49843b
    • A
      [PATCH] RTC subsystem: M48T86 driver · 1d98af87
      Alessandro Zummo 提交于
      Add a driver for the ST M48T86 / Dallas DS12887 RTC.
      
      This is a platform driver.  The platform device must provide I/O routines to
      access the RTC.
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Cc: Greg KH <greg@kroah.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      1d98af87
    • A
      [PATCH] RTC subsystem: I2C driver ids · f7f3682f
      Alessandro Zummo 提交于
      This patch adds the I2C driver ids to i2c-id.h in preparation of the I2C
      direct probing method.
      
      This is kept separate so that it can be integrated to
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      f7f3682f
    • A
      [PATCH] RTC subsystem: I2C cleanup · 6fc7f10c
      Alessandro Zummo 提交于
      This patch, completely optional, removes from drivers/i2c/chips all the
      drivers that are implemented in the new RTC subsystem.
      
      It should be noted that none of the current driver is actually integrated,
      i.e.  usable without further patches.
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Acked-by: NGreg Kroah-Hartman <gregkh@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      6fc7f10c
    • A
      [PATCH] RTC subsystem: class · 0c86edc0
      Alessandro Zummo 提交于
      Add the basic RTC subsystem infrastructure to the kernel.
      
      rtc/class.c - registration facilities for RTC drivers
      rtc/interface.c - kernel/rtc interface functions
      rtc/hctosys.c - snippet of code that copies hw clock to sw clock
      		at bootup, if configured to do so.
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Acked-by: NGreg Kroah-Hartman <gregkh@suse.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      0c86edc0
    • A
      [PATCH] RTC subsystem: ARM cleanup · 12b824fb
      Alessandro Zummo 提交于
      This patch removes from the ARM subsytem some of the rtc-related functions
      that have been included in the RTC subsystem.  It also fixes some naming
      collisions.
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      12b824fb
    • A
      [PATCH] RTC Subsystem: library functions · c58411e9
      Alessandro Zummo 提交于
      RTC and date/time related functions.
      Signed-off-by: NAlessandro Zummo <a.zummo@towertech.it>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      c58411e9
    • Y
      [PATCH] mips: fixed collision of rtc function name · d23ee8fe
      Yoichi Yuasa 提交于
      Fix the collision of rtc function name.
      Signed-off-by: NYoichi Yuasa <yoichi_yuasa@tripeaks.co.jp>
      Cc: Alessandro Zummo <a.zummo@towertech.it>
      Cc: Ralf Baechle <ralf@linux-mips.org>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      d23ee8fe
    • A
      [PATCH] Notifier chain update: API changes · e041c683
      Alan Stern 提交于
      The kernel's implementation of notifier chains is unsafe.  There is no
      protection against entries being added to or removed from a chain while the
      chain is in use.  The issues were discussed in this thread:
      
          http://marc.theaimsgroup.com/?l=linux-kernel&m=113018709002036&w=2
      
      We noticed that notifier chains in the kernel fall into two basic usage
      classes:
      
      	"Blocking" chains are always called from a process context
      	and the callout routines are allowed to sleep;
      
      	"Atomic" chains can be called from an atomic context and
      	the callout routines are not allowed to sleep.
      
      We decided to codify this distinction and make it part of the API.  Therefore
      this set of patches introduces three new, parallel APIs: one for blocking
      notifiers, one for atomic notifiers, and one for "raw" notifiers (which is
      really just the old API under a new name).  New kinds of data structures are
      used for the heads of the chains, and new routines are defined for
      registration, unregistration, and calling a chain.  The three APIs are
      explained in include/linux/notifier.h and their implementation is in
      kernel/sys.c.
      
      With atomic and blocking chains, the implementation guarantees that the chain
      links will not be corrupted and that chain callers will not get messed up by
      entries being added or removed.  For raw chains the implementation provides no
      guarantees at all; users of this API must provide their own protections.  (The
      idea was that situations may come up where the assumptions of the atomic and
      blocking APIs are not appropriate, so it should be possible for users to
      handle these things in their own way.)
      
      There are some limitations, which should not be too hard to live with.  For
      atomic/blocking chains, registration and unregistration must always be done in
      a process context since the chain is protected by a mutex/rwsem.  Also, a
      callout routine for a non-raw chain must not try to register or unregister
      entries on its own chain.  (This did happen in a couple of places and the code
      had to be changed to avoid it.)
      
      Since atomic chains may be called from within an NMI handler, they cannot use
      spinlocks for synchronization.  Instead we use RCU.  The overhead falls almost
      entirely in the unregister routine, which is okay since unregistration is much
      less frequent that calling a chain.
      
      Here is the list of chains that we adjusted and their classifications.  None
      of them use the raw API, so for the moment it is only a placeholder.
      
        ATOMIC CHAINS
        -------------
      arch/i386/kernel/traps.c:		i386die_chain
      arch/ia64/kernel/traps.c:		ia64die_chain
      arch/powerpc/kernel/traps.c:		powerpc_die_chain
      arch/sparc64/kernel/traps.c:		sparc64die_chain
      arch/x86_64/kernel/traps.c:		die_chain
      drivers/char/ipmi/ipmi_si_intf.c:	xaction_notifier_list
      kernel/panic.c:				panic_notifier_list
      kernel/profile.c:			task_free_notifier
      net/bluetooth/hci_core.c:		hci_notifier
      net/ipv4/netfilter/ip_conntrack_core.c:	ip_conntrack_chain
      net/ipv4/netfilter/ip_conntrack_core.c:	ip_conntrack_expect_chain
      net/ipv6/addrconf.c:			inet6addr_chain
      net/netfilter/nf_conntrack_core.c:	nf_conntrack_chain
      net/netfilter/nf_conntrack_core.c:	nf_conntrack_expect_chain
      net/netlink/af_netlink.c:		netlink_chain
      
        BLOCKING CHAINS
        ---------------
      arch/powerpc/platforms/pseries/reconfig.c:	pSeries_reconfig_chain
      arch/s390/kernel/process.c:		idle_chain
      arch/x86_64/kernel/process.c		idle_notifier
      drivers/base/memory.c:			memory_chain
      drivers/cpufreq/cpufreq.c		cpufreq_policy_notifier_list
      drivers/cpufreq/cpufreq.c		cpufreq_transition_notifier_list
      drivers/macintosh/adb.c:		adb_client_list
      drivers/macintosh/via-pmu.c		sleep_notifier_list
      drivers/macintosh/via-pmu68k.c		sleep_notifier_list
      drivers/macintosh/windfarm_core.c	wf_client_list
      drivers/usb/core/notify.c		usb_notifier_list
      drivers/video/fbmem.c			fb_notifier_list
      kernel/cpu.c				cpu_chain
      kernel/module.c				module_notify_list
      kernel/profile.c			munmap_notifier
      kernel/profile.c			task_exit_notifier
      kernel/sys.c				reboot_notifier_list
      net/core/dev.c				netdev_chain
      net/decnet/dn_dev.c:			dnaddr_chain
      net/ipv4/devinet.c:			inetaddr_chain
      
      It's possible that some of these classifications are wrong.  If they are,
      please let us know or submit a patch to fix them.  Note that any chain that
      gets called very frequently should be atomic, because the rwsem read-locking
      used for blocking chains is very likely to incur cache misses on SMP systems.
      (However, if the chain's callout routines may sleep then the chain cannot be
      atomic.)
      
      The patch set was written by Alan Stern and Chandra Seetharaman, incorporating
      material written by Keith Owens and suggestions from Paul McKenney and Andrew
      Morton.
      
      [jes@sgi.com: restructure the notifier chain initialization macros]
      Signed-off-by: NAlan Stern <stern@rowland.harvard.edu>
      Signed-off-by: NChandra Seetharaman <sekharan@us.ibm.com>
      Signed-off-by: NJes Sorensen <jes@sgi.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      e041c683
    • I
      [PATCH] lightweight robust futexes updates 2 · 76b81e2b
      Ingo Molnar 提交于
      futex.h updates:
      
      - get rid of FUTEX_OWNER_PENDING - it's not used
      - reduce ROBUST_LIST_LIMIT to a saner value
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      76b81e2b
    • I
      [PATCH] lightweight robust futexes updates · 8f17d3a5
      Ingo Molnar 提交于
      - fix: initialize the robust list(s) to NULL in copy_process.
      
      - doc update
      
      - cleanup: rename _inuser to _inatomic
      
      - __user cleanups and other small cleanups
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Arjan van de Ven <arjan@infradead.org>
      Cc: Ulrich Drepper <drepper@redhat.com>
      Cc: Andi Kleen <ak@muc.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      8f17d3a5
    • I
      [PATCH] lightweight robust futexes: x86_64 · 8fdd6c6d
      Ingo Molnar 提交于
      x86_64: add the futex_atomic_cmpxchg_inuser() assembly implementation, and
      wire up the new syscalls.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Signed-off-by: NArjan van de Ven <arjan@infradead.org>
      Acked-by: NUlrich Drepper <drepper@redhat.com>
      Cc: Andi Kleen <ak@muc.de>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      8fdd6c6d
    • I
      [PATCH] lightweight robust futexes: i386 · dfd4e3ec
      Ingo Molnar 提交于
      i386: add the futex_atomic_cmpxchg_inuser() assembly implementation, and wire
      up the new syscalls.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Signed-off-by: NArjan van de Ven <arjan@infradead.org>
      Acked-by: NUlrich Drepper <drepper@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      dfd4e3ec
    • I
      [PATCH] lightweight robust futexes: compat · 34f192c6
      Ingo Molnar 提交于
      32-bit syscall compatibility support.  (This patch also moves all futex
      related compat functionality into kernel/futex_compat.c.)
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Signed-off-by: NArjan van de Ven <arjan@infradead.org>
      Acked-by: NUlrich Drepper <drepper@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      34f192c6
    • I
      [PATCH] lightweight robust futexes: core · 0771dfef
      Ingo Molnar 提交于
      Add the core infrastructure for robust futexes: structure definitions, the new
      syscalls and the do_exit() based cleanup mechanism.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Signed-off-by: NArjan van de Ven <arjan@infradead.org>
      Acked-by: NUlrich Drepper <drepper@redhat.com>
      Cc: Michael Kerrisk <mtk-manpages@gmx.net>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      0771dfef
    • I
      [PATCH] lightweight robust futexes: arch defaults · e9056f13
      Ingo Molnar 提交于
      This patchset provides a new (written from scratch) implementation of robust
      futexes, called "lightweight robust futexes".  We believe this new
      implementation is faster and simpler than the vma-based robust futex solutions
      presented before, and we'd like this patchset to be adopted in the upstream
      kernel.  This is version 1 of the patchset.
      
        Background
        ----------
      
      What are robust futexes?  To answer that, we first need to understand what
      futexes are: normal futexes are special types of locks that in the
      noncontended case can be acquired/released from userspace without having to
      enter the kernel.
      
      A futex is in essence a user-space address, e.g.  a 32-bit lock variable
      field.  If userspace notices contention (the lock is already owned and someone
      else wants to grab it too) then the lock is marked with a value that says
      "there's a waiter pending", and the sys_futex(FUTEX_WAIT) syscall is used to
      wait for the other guy to release it.  The kernel creates a 'futex queue'
      internally, so that it can later on match up the waiter with the waker -
      without them having to know about each other.  When the owner thread releases
      the futex, it notices (via the variable value) that there were waiter(s)
      pending, and does the sys_futex(FUTEX_WAKE) syscall to wake them up.  Once all
      waiters have taken and released the lock, the futex is again back to
      'uncontended' state, and there's no in-kernel state associated with it.  The
      kernel completely forgets that there ever was a futex at that address.  This
      method makes futexes very lightweight and scalable.
      
      "Robustness" is about dealing with crashes while holding a lock: if a process
      exits prematurely while holding a pthread_mutex_t lock that is also shared
      with some other process (e.g.  yum segfaults while holding a pthread_mutex_t,
      or yum is kill -9-ed), then waiters for that lock need to be notified that the
      last owner of the lock exited in some irregular way.
      
      To solve such types of problems, "robust mutex" userspace APIs were created:
      pthread_mutex_lock() returns an error value if the owner exits prematurely -
      and the new owner can decide whether the data protected by the lock can be
      recovered safely.
      
      There is a big conceptual problem with futex based mutexes though: it is the
      kernel that destroys the owner task (e.g.  due to a SEGFAULT), but the kernel
      cannot help with the cleanup: if there is no 'futex queue' (and in most cases
      there is none, futexes being fast lightweight locks) then the kernel has no
      information to clean up after the held lock!  Userspace has no chance to clean
      up after the lock either - userspace is the one that crashes, so it has no
      opportunity to clean up.  Catch-22.
      
      In practice, when e.g.  yum is kill -9-ed (or segfaults), a system reboot is
      needed to release that futex based lock.  This is one of the leading
      bugreports against yum.
      
      To solve this problem, 'Robust Futex' patches were created and presented on
      lkml: the one written by Todd Kneisel and David Singleton is the most advanced
      at the moment.  These patches all tried to extend the futex abstraction by
      registering futex-based locks in the kernel - and thus give the kernel a
      chance to clean up.
      
      E.g.  in David Singleton's robust-futex-6.patch, there are 3 new syscall
      variants to sys_futex(): FUTEX_REGISTER, FUTEX_DEREGISTER and FUTEX_RECOVER.
      The kernel attaches such robust futexes to vmas (via
      vma->vm_file->f_mapping->robust_head), and at do_exit() time, all vmas are
      searched to see whether they have a robust_head set.
      
      Lots of work went into the vma-based robust-futex patch, and recently it has
      improved significantly, but unfortunately it still has two fundamental
      problems left:
      
       - they have quite complex locking and race scenarios.  The vma-based
         patches had been pending for years, but they are still not completely
         reliable.
      
       - they have to scan _every_ vma at sys_exit() time, per thread!
      
      The second disadvantage is a real killer: pthread_exit() takes around 1
      microsecond on Linux, but with thousands (or tens of thousands) of vmas every
      pthread_exit() takes a millisecond or more, also totally destroying the CPU's
      L1 and L2 caches!
      
      This is very much noticeable even for normal process sys_exit_group() calls:
      the kernel has to do the vma scanning unconditionally!  (this is because the
      kernel has no knowledge about how many robust futexes there are to be cleaned
      up, because a robust futex might have been registered in another task, and the
      futex variable might have been simply mmap()-ed into this process's address
      space).
      
      This huge overhead forced the creation of CONFIG_FUTEX_ROBUST, but worse than
      that: the overhead makes robust futexes impractical for any type of generic
      Linux distribution.
      
      So it became clear to us, something had to be done.  Last week, when Thomas
      Gleixner tried to fix up the vma-based robust futex patch in the -rt tree, he
      found a handful of new races and we were talking about it and were analyzing
      the situation.  At that point a fundamentally different solution occured to
      me.  This patchset (written in the past couple of days) implements that new
      solution.  Be warned though - the patchset does things we normally dont do in
      Linux, so some might find the approach disturbing.  Parental advice
      recommended ;-)
      
        New approach to robust futexes
        ------------------------------
      
      At the heart of this new approach there is a per-thread private list of robust
      locks that userspace is holding (maintained by glibc) - which userspace list
      is registered with the kernel via a new syscall [this registration happens at
      most once per thread lifetime].  At do_exit() time, the kernel checks this
      user-space list: are there any robust futex locks to be cleaned up?
      
      In the common case, at do_exit() time, there is no list registered, so the
      cost of robust futexes is just a simple current->robust_list != NULL
      comparison.  If the thread has registered a list, then normally the list is
      empty.  If the thread/process crashed or terminated in some incorrect way then
      the list might be non-empty: in this case the kernel carefully walks the list
      [not trusting it], and marks all locks that are owned by this thread with the
      FUTEX_OWNER_DEAD bit, and wakes up one waiter (if any).
      
      The list is guaranteed to be private and per-thread, so it's lockless.  There
      is one race possible though: since adding to and removing from the list is
      done after the futex is acquired by glibc, there is a few instructions window
      for the thread (or process) to die there, leaving the futex hung.  To protect
      against this possibility, userspace (glibc) also maintains a simple per-thread
      'list_op_pending' field, to allow the kernel to clean up if the thread dies
      after acquiring the lock, but just before it could have added itself to the
      list.  Glibc sets this list_op_pending field before it tries to acquire the
      futex, and clears it after the list-add (or list-remove) has finished.
      
      That's all that is needed - all the rest of robust-futex cleanup is done in
      userspace [just like with the previous patches].
      
      Ulrich Drepper has implemented the necessary glibc support for this new
      mechanism, which fully enables robust mutexes.  (Ulrich plans to commit these
      changes to glibc-HEAD later today.)
      
      Key differences of this userspace-list based approach, compared to the vma
      based method:
      
       - it's much, much faster: at thread exit time, there's no need to loop
         over every vma (!), which the VM-based method has to do.  Only a very
         simple 'is the list empty' op is done.
      
       - no VM changes are needed - 'struct address_space' is left alone.
      
       - no registration of individual locks is needed: robust mutexes dont need
         any extra per-lock syscalls.  Robust mutexes thus become a very lightweight
         primitive - so they dont force the application designer to do a hard choice
         between performance and robustness - robust mutexes are just as fast.
      
       - no per-lock kernel allocation happens.
      
       - no resource limits are needed.
      
       - no kernel-space recovery call (FUTEX_RECOVER) is needed.
      
       - the implementation and the locking is "obvious", and there are no
         interactions with the VM.
      
        Performance
        -----------
      
      I have benchmarked the time needed for the kernel to process a list of 1
      million (!) held locks, using the new method [on a 2GHz CPU]:
      
       - with FUTEX_WAIT set [contended mutex]: 130 msecs
       - without FUTEX_WAIT set [uncontended mutex]: 30 msecs
      
      I have also measured an approach where glibc does the lock notification [which
      it currently does for !pshared robust mutexes], and that took 256 msecs -
      clearly slower, due to the 1 million FUTEX_WAKE syscalls userspace had to do.
      
      (1 million held locks are unheard of - we expect at most a handful of locks to
      be held at a time.  Nevertheless it's nice to know that this approach scales
      nicely.)
      
        Implementation details
        ----------------------
      
      The patch adds two new syscalls: one to register the userspace list, and one
      to query the registered list pointer:
      
       asmlinkage long
       sys_set_robust_list(struct robust_list_head __user *head,
                           size_t len);
      
       asmlinkage long
       sys_get_robust_list(int pid, struct robust_list_head __user **head_ptr,
                           size_t __user *len_ptr);
      
      List registration is very fast: the pointer is simply stored in
      current->robust_list.  [Note that in the future, if robust futexes become
      widespread, we could extend sys_clone() to register a robust-list head for new
      threads, without the need of another syscall.]
      
      So there is virtually zero overhead for tasks not using robust futexes, and
      even for robust futex users, there is only one extra syscall per thread
      lifetime, and the cleanup operation, if it happens, is fast and
      straightforward.  The kernel doesnt have any internal distinction between
      robust and normal futexes.
      
      If a futex is found to be held at exit time, the kernel sets the highest bit
      of the futex word:
      
      	#define FUTEX_OWNER_DIED        0x40000000
      
      and wakes up the next futex waiter (if any). User-space does the rest of
      the cleanup.
      
      Otherwise, robust futexes are acquired by glibc by putting the TID into the
      futex field atomically.  Waiters set the FUTEX_WAITERS bit:
      
      	#define FUTEX_WAITERS           0x80000000
      
      and the remaining bits are for the TID.
      
        Testing, architecture support
        -----------------------------
      
      I've tested the new syscalls on x86 and x86_64, and have made sure the parsing
      of the userspace list is robust [ ;-) ] even if the list is deliberately
      corrupted.
      
      i386 and x86_64 syscalls are wired up at the moment, and Ulrich has tested the
      new glibc code (on x86_64 and i386), and it works for his robust-mutex
      testcases.
      
      All other architectures should build just fine too - but they wont have the
      new syscalls yet.
      
      Architectures need to implement the new futex_atomic_cmpxchg_inuser() inline
      function before writing up the syscalls (that function returns -ENOSYS right
      now).
      
      This patch:
      
      Add placeholder futex_atomic_cmpxchg_inuser() implementations to every
      architecture that supports futexes.  It returns -ENOSYS.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Signed-off-by: NArjan van de Ven <arjan@infradead.org>
      Acked-by: NUlrich Drepper <drepper@redhat.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      e9056f13
    • I
      [PATCH] mips: add ptr_to_compat() · 62ac285f
      Ingo Molnar 提交于
      Add ptr_to_compat() - needed by the new robust futex code.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Cc: Ralf Baechle <ralf@linux-mips.org>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      62ac285f
    • I
      [PATCH] parisc: add ptr_to_compat() · 213b63b7
      Ingo Molnar 提交于
      Add ptr_to_compat() to parisc - needed by the new robust futex code.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Cc: Kyle McMartin <kyle@mcmartin.ca>
      Cc: Grant Grundler <iod00d@hp.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      213b63b7
    • I
      [PATCH] s390: add ptr_to_compat() · f267fa9f
      Ingo Molnar 提交于
      Add ptr_to_compat() to s390 - needed by the new robust-futex code.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
      Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
      
      untested. CHECKME: am i right about the 0x7fffffffUL masking?
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      f267fa9f
    • I
      [PATCH] ia64: add ptr_to_compat() · 66e863ac
      Ingo Molnar 提交于
      Add ptr_to_compat() to ia64 - needed by the robust-futex code.
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Cc: "Luck, Tony" <tony.luck@intel.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      66e863ac
    • D
      [PATCH] unify PFN_* macros · 22a9835c
      Dave Hansen 提交于
      Just about every architecture defines some macros to do operations on pfns.
       They're all virtually identical.  This patch consolidates all of them.
      
      One minor glitch is that at least i386 uses them in a very skeletal header
      file.  To keep away from #include dependency hell, I stuck the new
      definitions in a new, isolated header.
      
      Of all of the implementations, sh64 is the only one that varied by a bit.
      It used some masks to ensure that any sign-extension got ripped away before
      the arithmetic is done.  This has been posted to that sh64 maintainers and
      the development list.
      
      Compiles on x86, x86_64, ia64 and ppc64.
      Signed-off-by: NDave Hansen <haveblue@us.ibm.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      22a9835c
    • K
      [PATCH] uninline zone helpers · 95144c78
      KAMEZAWA Hiroyuki 提交于
      Helper functions for for_each_online_pgdat/for_each_zone look too big to be
      inlined.  Speed of these helper macro itself is not very important.  (inner
      loops are tend to do more work than this)
      
      This patch make helper function to be out-of-lined.
      
      	inline		out-of-line
      .text   005c0680        005bf6a0
      
      005c0680 - 005bf6a0 = FE0 = 4Kbytes.
      Signed-off-by: NKAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      95144c78
    • K
      [PATCH] for_each_online_pgdat: remove pgdat_list · ae0f15fb
      KAMEZAWA Hiroyuki 提交于
      By using for_each_online_pgdat(), pgdat_list is not necessary now.  This patch
      removes it.
      Signed-off-by: NKAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      ae0f15fb
    • K
      [PATCH] for_each_online_pgdat: for_each_bootmem · 679bc9fb
      KAMEZAWA Hiroyuki 提交于
      Add a list_head to bootmem_data_t and make bootmems use it.  bootmem list is
      sorted by node_boot_start.
      
      Only nodes against which init_bootmem() is called are linked to the list.
      (i386 allocates bootmem only from one node(0) not from all online nodes.)
      
      A summary:
       1. for_each_online_pgdat() traverses all *online* nodes.
       2. alloc_bootmem() allocates memory only from initialized-for-bootmem nodes.
      Signed-off-by: NKAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      679bc9fb