1. 18 2月, 2008 4 次提交
  2. 14 2月, 2008 8 次提交
  3. 13 2月, 2008 1 次提交
  4. 10 2月, 2008 3 次提交
  5. 09 2月, 2008 13 次提交
  6. 08 2月, 2008 11 次提交
    • J
      Enhanced partition statistics: documentation update · 0e53c2be
      Jerome Marchand 提交于
      Update the documentation to reflect the change in userspace interface.
      Signed-off-by: NJerome Marchand <jmarchan@redhat.com>
      Signed-off-by: NJens Axboe <jens.axboe@oracle.com>
      0e53c2be
    • L
      ACPI: thermal: syntax, spelling, kernel-doc · 543a9561
      Len Brown 提交于
      Reviewed-by: NRandy Dunlap <randy.dunlap@oracle.com>
      Signed-off-by: NLen Brown <len.brown@intel.com>
      543a9561
    • C
      SLUB: Support for performance statistics · 8ff12cfc
      Christoph Lameter 提交于
      The statistics provided here allow the monitoring of allocator behavior but
      at the cost of some (minimal) loss of performance. Counters are placed in
      SLUB's per cpu data structure. The per cpu structure may be extended by the
      statistics to grow larger than one cacheline which will increase the cache
      footprint of SLUB.
      
      There is a compile option to enable/disable the inclusion of the runtime
      statistics and its off by default.
      
      The slabinfo tool is enhanced to support these statistics via two options:
      
      -D 	Switches the line of information displayed for a slab from size
      	mode to activity mode.
      
      -A	Sorts the slabs displayed by activity. This allows the display of
      	the slabs most important to the performance of a certain load.
      
      -r	Report option will report detailed statistics on
      
      Example (tbench load):
      
      slabinfo -AD		->Shows the most active slabs
      
      Name                   Objects    Alloc     Free   %Fast
      skbuff_fclone_cache         33 111953835 111953835  99  99
      :0000192                  2666  5283688  5281047  99  99
      :0001024                   849  5247230  5246389  83  83
      vm_area_struct            1349   119642   118355  91  22
      :0004096                    15    66753    66751  98  98
      :0000064                  2067    25297    23383  98  78
      dentry                   10259    28635    18464  91  45
      :0000080                 11004    18950     8089  98  98
      :0000096                  1703    12358    10784  99  98
      :0000128                   762    10582     9875  94  18
      :0000512                   184     9807     9647  95  81
      :0002048                   479     9669     9195  83  65
      anon_vma                   777     9461     9002  99  71
      kmalloc-8                 6492     9981     5624  99  97
      :0000768                   258     7174     6931  58  15
      
      So the skbuff_fclone_cache is of highest importance for the tbench load.
      Pretty high load on the 192 sized slab. Look for the aliases
      
      slabinfo -a | grep 000192
      :0000192     <- xfs_btree_cur filp kmalloc-192 uid_cache tw_sock_TCP
      	request_sock_TCPv6 tw_sock_TCPv6 skbuff_head_cache xfs_ili
      
      Likely skbuff_head_cache.
      
      
      Looking into the statistics of the skbuff_fclone_cache is possible through
      
      slabinfo skbuff_fclone_cache	->-r option implied if cache name is mentioned
      
      
      .... Usual output ...
      
      Slab Perf Counter       Alloc     Free %Al %Fr
      --------------------------------------------------
      Fastpath             111953360 111946981  99  99
      Slowpath                 1044     7423   0   0
      Page Alloc                272      264   0   0
      Add partial                25      325   0   0
      Remove partial             86      264   0   0
      RemoteObj/SlabFrozen      350     4832   0   0
      Total                111954404 111954404
      
      Flushes       49 Refill        0
      Deactivate Full=325(92%) Empty=0(0%) ToHead=24(6%) ToTail=1(0%)
      
      Looks good because the fastpath is overwhelmingly taken.
      
      
      skbuff_head_cache:
      
      Slab Perf Counter       Alloc     Free %Al %Fr
      --------------------------------------------------
      Fastpath              5297262  5259882  99  99
      Slowpath                 4477    39586   0   0
      Page Alloc                937      824   0   0
      Add partial                 0     2515   0   0
      Remove partial           1691      824   0   0
      RemoteObj/SlabFrozen     2621     9684   0   0
      Total                 5301739  5299468
      
      Deactivate Full=2620(100%) Empty=0(0%) ToHead=0(0%) ToTail=0(0%)
      
      
      Descriptions of the output:
      
      Total:		The total number of allocation and frees that occurred for a
      		slab
      
      Fastpath:	The number of allocations/frees that used the fastpath.
      
      Slowpath:	Other allocations
      
      Page Alloc:	Number of calls to the page allocator as a result of slowpath
      		processing
      
      Add Partial:	Number of slabs added to the partial list through free or
      		alloc (occurs during cpuslab flushes)
      
      Remove Partial:	Number of slabs removed from the partial list as a result of
      		allocations retrieving a partial slab or by a free freeing
      		the last object of a slab.
      
      RemoteObj/Froz:	How many times were remotely freed object encountered when a
      		slab was about to be deactivated. Frozen: How many times was
      		free able to skip list processing because the slab was in use
      		as the cpuslab of another processor.
      
      Flushes:	Number of times the cpuslab was flushed on request
      		(kmem_cache_shrink, may result from races in __slab_alloc)
      
      Refill:		Number of times we were able to refill the cpuslab from
      		remotely freed objects for the same slab.
      
      Deactivate:	Statistics how slabs were deactivated. Shows how they were
      		put onto the partial list.
      
      In general fastpath is very good. Slowpath without partial list processing is
      also desirable. Any touching of partial list uses node specific locks which
      may potentially cause list lock contention.
      Signed-off-by: NChristoph Lameter <clameter@sgi.com>
      8ff12cfc
    • J
      hwmon: (w83627ehf) The W83627DHG has 8 VID pins · cbe311f2
      Jean Delvare 提交于
      While the W83627EHF/EHG has only 6 VID pins, the W83627DHG has 8 VID
      pins, to support VRD 11.0. Add support for this.
      Signed-off-by: NJean Delvare <khali@linux-fr.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      cbe311f2
    • S
    • J
      hwmon: (w83781d) Drop W83627HF support · 05663368
      Jean Delvare 提交于
      The W83627HF hardware monitoring features are supported by the
      w83627hf driver for several years now. Support by the w83781d has
      been advertised as deprecated 6 months ago, it's about time to see
      it go.
      Signed-off-by: NJean Delvare <khali@linux-fr.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      05663368
    • J
      f1d8e332
    • J
      hwmon: Update the lm-sensors website address · ec1d86c4
      Jean Delvare 提交于
      It's about time to reflect the move of the lm-sensors project to
      lm-sensors.org.
      Signed-off-by: NJean Delvare <khali@linux-fr.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      ec1d86c4
    • J
      hwmon: (lm87) Add support for the Analog Devices ADM1024 · c7fa3737
      Jean Delvare 提交于
      It happens that the Analog Devices ADM1024 is fully compatible with
      the National Semiconductor LM87, so support for the former can easily
      be added to the lm87 driver.
      Signed-off-by: NJean Delvare <khali@linux-fr.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      c7fa3737
    • K
      hwmon: Add support for Winbond W83L786NG/NR · 85f03bcc
      Kevin Lo 提交于
      Signed-off-by: NKevin Lo <kevlo@kevlo.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      85f03bcc
    • J
      hwmon: (lm78/w83781d) Probe fewer I2C addresses · 6722fead
      Jean Delvare 提交于
      We've never seen any device supported by the lm78 or w83781d driver at
      addresses 0x20-0x27, so let's stop probing these addresses. Extra probes cost
      time, and have potential for confusing or misdetecting other I2C devices.
      Signed-off-by: NJean Delvare <khali@linux-fr.org>
      Signed-off-by: NMark M. Hoffman <mhoffman@lightlink.com>
      6722fead