1. 27 6月, 2014 2 次提交
  2. 23 6月, 2014 6 次提交
  3. 19 6月, 2014 8 次提交
  4. 16 6月, 2014 1 次提交
  5. 06 6月, 2014 4 次提交
  6. 21 5月, 2014 1 次提交
  7. 19 5月, 2014 1 次提交
    • P
      block: optimize zero writes with bdrv_write_zeroes · 465bee1d
      Peter Lieven 提交于
      this patch tries to optimize zero write requests
      by automatically using bdrv_write_zeroes if it is
      supported by the format.
      
      This significantly speeds up file system initialization and
      should speed zero write test used to test backend storage
      performance.
      
      I ran the following 2 tests on my internal SSD with a
      50G QCOW2 container and on an attached iSCSI storage.
      
      a) mkfs.ext4 -E lazy_itable_init=0,lazy_journal_init=0 /dev/vdX
      
      QCOW2         [off]     [on]     [unmap]
      -----
      runtime:       14secs    1.1secs  1.1secs
      filesize:      937M      18M      18M
      
      iSCSI         [off]     [on]     [unmap]
      ----
      runtime:       9.3s      0.9s     0.9s
      
      b) dd if=/dev/zero of=/dev/vdX bs=1M oflag=direct
      
      QCOW2         [off]     [on]     [unmap]
      -----
      runtime:       246secs   18secs   18secs
      filesize:      51G       192K     192K
      throughput:    203M/s    2.3G/s   2.3G/s
      
      iSCSI*        [off]     [on]     [unmap]
      ----
      runtime:       8mins     45secs   33secs
      throughput:    106M/s    1.2G/s   1.6G/s
      allocated:     100%      100%     0%
      
      * The storage was connected via an 1Gbit interface.
        It seems to internally handle writing zeroes
        via WRITESAME16 very fast.
      Signed-off-by: NPeter Lieven <pl@kamp.de>
      Signed-off-by: NKevin Wolf <kwolf@redhat.com>
      465bee1d
  8. 06 5月, 2014 2 次提交
  9. 26 4月, 2014 1 次提交
  10. 26 3月, 2014 1 次提交
  11. 13 3月, 2014 1 次提交
  12. 11 3月, 2014 1 次提交
  13. 05 3月, 2014 3 次提交
  14. 01 3月, 2014 3 次提交
  15. 25 2月, 2014 1 次提交
  16. 22 2月, 2014 1 次提交
  17. 18 2月, 2014 1 次提交
  18. 15 2月, 2014 1 次提交
  19. 09 2月, 2014 1 次提交
    • P
      block: add native support for NFS · 6542aa9c
      Peter Lieven 提交于
      This patch adds native support for accessing images on NFS
      shares without the requirement to actually mount the entire
      NFS share on the host.
      
      NFS Images can simply be specified by an url of the form:
      nfs://<host>/<export>/<filename>[?param=value[&param2=value2[&...]]]
      
      For example:
      qemu-img create -f qcow2 nfs://10.0.0.1/qemu-images/test.qcow2
      
      You need LibNFS from Ronnie Sahlberg available at:
         git://github.com/sahlberg/libnfs.git
      for this to work.
      
      During configure it is automatically probed for libnfs and support
      is enabled on-the-fly. You can forbid or enforce libnfs support
      with --disable-libnfs or --enable-libnfs respectively.
      
      Due to NFS restrictions you might need to execute your binaries
      as root, allow them to open priviledged ports (<1024) or specify
      insecure option on the NFS server.
      
      For additional information on ROOT vs. non-ROOT operation and URL
      format + parameters see:
         https://raw.github.com/sahlberg/libnfs/master/README
      
      Supported by qemu are the uid, gid and tcp-syncnt URL parameters.
      
      LibNFS currently support NFS version 3 only.
      Signed-off-by: NPeter Lieven <pl@kamp.de>
      Signed-off-by: NKevin Wolf <kwolf@redhat.com>
      6542aa9c