1. 12 2月, 2015 1 次提交
  2. 29 1月, 2015 1 次提交
    • M
      qemu: Allow UEFI paths to be specified at compile time · bc03a231
      Michal Privoznik 提交于
      Up until now there are just two ways how to specify UEFI paths to
      libvirt. The first one is editing qemu.conf, the other is editing
      qemu_conf.c and recompile which is not that fancy. So, new
      configure option is introduced: --with-loader-nvram which takes a
      list of pairs of UEFI firmware and NVRAM store. This way, the
      compiled in defaults can be passed during compile time without
      need to change the code itself.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      bc03a231
  3. 27 1月, 2015 2 次提交
  4. 14 1月, 2015 1 次提交
    • J
      Introduce support for parsing/formatting Xen xl config format · 4689cdf7
      Jim Fehlig 提交于
      Introduce a parser/formatter for the xl config format.  Since the
      deprecation of xm/xend, the VM config file format has diverged as
      new features are added to libxl.  This patch adds support for parsing
      and formating the xl config format.  It supports the existing xm config
      format, plus adds support for spice graphics and xl disk config syntax.
      
      Disk config is specified a bit differently in xl as compared to xm.  In
      xl, disk config consists of comma-separated positional parameters and
      keyword/value pairs separated by commas. Positional parameters are
      specified as follows
      
         target, format, vdev, access
      
      Supported keys for key=value options are
      
        devtype, backendtype
      
      The positional paramters can also be specified in key/value form.  For
      example the following xl disk config are equivalent
      
          /dev/vg/guest-volume,,hda
          /dev/vg/guest-volume,raw,hda,rw
          format=raw, vdev=hda, access=rw, target=/dev/vg/guest-volume
      
      See $xen_sources/docs/misc/xl-disk-configuration.txt for more details.
      
      xl disk config is parsed with the help of xlu_disk_parse() from
      libxlutil, libxl's utility library.  Although the library exists
      in all Xen versions supported by the libxl virt driver, only
      recently has the corresponding header file been included.  A check
      for the header is done in configure.ac.  If not found, xlu_disk_parse()
      is declared externally.
      Signed-off-by: NKiarie Kahurani <davidkiarie4@gmail.com>
      Signed-off-by: NJim Fehlig <jfehlig@suse.com>
      4689cdf7
  5. 13 1月, 2015 1 次提交
  6. 04 1月, 2015 1 次提交
    • K
      src/xenconfig: Xen-xl parser · 2c78051a
      Kiarie Kahurani 提交于
      Introduce a Xen xl parser
      
      This parser allows for users to convert the new xl disk format and
      spice graphics config to libvirt xml format and vice versa. Regarding
      the spice graphics config, the code is pretty much straight forward.
      For the disk {formating, parsing}, this parser takes care of the new
      xl format which include positional parameters and key/value parameters.
      In xl format disk config a <diskspec> consists of parameters separated by
      commas. If the parameters do not contain an '=' they are automatically
      assigned to certain options following the order below
      
         target, format, vdev, access
      
      The above are the only mandatory parameters in the <diskspec> but there
      are many more disk config options. These options can be specified as
      key=value pairs. This takes care of the rest of the options such as
      
        devtype, backend, backendtype, script, direct-io-safe,
      
      The positional paramters can also be specified in key/value form
      for example
      
          /dev/vg/guest-volume,,hda
          /dev/vg/guest-volume,raw,hda,rw
          format=raw, vdev=hda, access=rw, target=/dev/vg/guest-volume
      
      are interpleted to one config.
      
      In xm format, the above diskspec would be written as
      
      phy:/dev/vg/guest-volume,hda,w
      
      The disk parser is based on the same parser used successfully by
      the Xen project for several years now.  Ian Jackson authored the
      scanner, which is used by this commit with mimimal changes.  Only
      the PREFIX option is changed, to produce function and file names
      more consistent with libvirt's convention.
      Signed-off-by: NKiarie Kahurani <davidkiarie4@gmail.com>
      Signed-off-by: NJim Fehlig <jfehlig@suse.com>
      2c78051a
  7. 13 12月, 2014 1 次提交
  8. 20 11月, 2014 1 次提交
  9. 03 11月, 2014 1 次提交
  10. 30 10月, 2014 1 次提交
    • R
      virnetdev: stub virNetDev{Add,Del}Multi on FreeBSD · 00fa136d
      Roman Bogorodskiy 提交于
      Currently, build fails on FreeBSD because its struct ifreq does not
      have ifr_hwaddr member. In order to fix that, check if this member
      is present, otherwise fall back to the stub version of the
      virNetDev{Add,Del}Multi functions.
      00fa136d
  11. 03 10月, 2014 1 次提交
  12. 02 10月, 2014 1 次提交
  13. 01 10月, 2014 1 次提交
  14. 25 9月, 2014 1 次提交
    • D
      parallels: build with parallels SDK · 64018e0c
      Dmitry Guryanov 提交于
      Executing prlctl command is not an optimal way to interact with
      Parallels Cloud Server (PCS), it's better to use parallels SDK,
      which is a remote API to paralles dispatcher service.
      
      We prepared opensource version of this SDK and published it on
      github, it's distributed under LGPL license. Here is a git repo:
      https://github.com/Parallels/parallels-sdk.
      
      To build with parallels SDK user should get compiler and linker
      options from pkg-config 'parallels-sdk' file. So fix checks in
      configure script and build with parallels SDK, if that pkg-config
      file exists and add gcc options to makefile.
      Signed-off-by: NDmitry Guryanov <dguryanov@parallels.com>
      64018e0c
  15. 02 9月, 2014 1 次提交
  16. 19 8月, 2014 1 次提交
  17. 14 8月, 2014 1 次提交
  18. 12 8月, 2014 1 次提交
    • R
      storage: ZFS support · 0257d06b
      Roman Bogorodskiy 提交于
      Implement ZFS storage backend driver. Currently supported
      only on FreeBSD because of ZFS limitations on Linux.
      
      Features supported:
      
       - pool-start, pool-stop
       - pool-info
       - vol-list
       - vol-create / vol-delete
      
      Pool definition looks like that:
      
       <pool type='zfs'>
        <name>myzfspool</name>
        <source>
          <name>actualpoolname</name>
        </source>
       </pool>
      
      The 'actualpoolname' value is a name of the pool on the system,
      such as shown by 'zpool list' command. Target makes no sense
      here because volumes path is always /dev/zvol/$poolname/$volname.
      
      User has to create a pool on his own, this driver doesn't
      support pool creation currently.
      
      A volume could be used with Qemu by adding an entry like this:
      
          <disk type='volume' device='disk'>
            <driver name='qemu' type='raw'/>
            <source pool='myzfspool' volume='vol5'/>
            <target dev='hdc' bus='ide'/>
          </disk>
      0257d06b
  19. 04 8月, 2014 1 次提交
  20. 18 7月, 2014 1 次提交
    • M
      examples: Introduce domtop · 28d54aab
      Michal Privoznik 提交于
      There's this question on the list that is asked over and over again.
      How do I get {cpu, memory, ...} usage in percentage? Or its modified
      version: How do I plot nice graphs like virt-manager does?
      
      It would be nice if we have an example to inspire people. And that's
      what domtop should do. Yes, it could be written in different ways, but
      I've chosen this one as I think it show explicitly what users need to
      implement in order to imitate virt-manager's graphing.
      
      Note: The usage is displayed from host perspective. That is, how much
      host CPUs the domain is using. But it should be fairly simple to
      switch do just guest CPU usage if needed.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      28d54aab
  21. 16 7月, 2014 1 次提交
  22. 02 7月, 2014 1 次提交
  23. 23 6月, 2014 2 次提交
  24. 18 6月, 2014 1 次提交
  25. 02 6月, 2014 1 次提交
  26. 05 5月, 2014 1 次提交
    • R
      bhyve: report cpuTime in bhyveDomainGetInfo · 0541727c
      Roman Bogorodskiy 提交于
      Add a helper function virBhyveGetDomainTotalCpuStats() to
      obtain process CPU time using kvm (kernel memory interface)
      and use it to set cpuTime field of the virDomainInfo struct in
      bhyveDomainGetInfo().
      0541727c
  27. 04 5月, 2014 1 次提交
  28. 30 4月, 2014 1 次提交
  29. 29 4月, 2014 1 次提交
  30. 12 4月, 2014 1 次提交
  31. 07 4月, 2014 1 次提交
  32. 01 4月, 2014 2 次提交
  33. 21 3月, 2014 2 次提交
    • J
      Fix typo in configure.ac · b7a5a6ab
      Ján Tomko 提交于
      s/profram/program/
      b7a5a6ab
    • J
      Define ISCSIADM unconditionally · 69ad6b57
      Ján Tomko 提交于
      Commit 5e1d5dde moved the functions using isciadm to util/viriscsi.c,
      but ISCSIADM was only defined when the iscsi backend was compiled in.
      
      Define it unconditionally to fix the build (and allow testing
      of viriscsi.c) even if iscsi backend is disabled.
      69ad6b57
  34. 06 3月, 2014 1 次提交
    • R
      Fix missing char dev lock path case in configure · 2c4870d0
      Roman Bogorodskiy 提交于
      configure check for character devices lock path calls
      AC_DEFINE_UNQUOTED for VIR_CHRDEV_LOCK_FILE_PATH even if
      $with_chrdev_lock_files = "no".
      
      So the locking code in conf/virchrdev.c:
      
       #ifdef VIR_CHRDEV_LOCK_FILE_PATH
      
      is compiled in even if it shouldn't, because VIR_CHRDEV_LOCK_FILE_PATH
      is defined as "no", so it tries to create lock files with strange
      lock path like 'no/LCK..'.
      
      Fix that by calling AC_DEFINE_UNQUOTED only if $with_chrdev_lock_files
      is not 'no'.
      2c4870d0
  35. 03 3月, 2014 1 次提交
  36. 24 2月, 2014 1 次提交
    • M
      virNetServerRun: Notify systemd that we're accepting clients · 68954fb2
      Michal Privoznik 提交于
      Systemd does not forget about the cases, where client service needs to
      wait for daemon service to initialize and start accepting new clients.
      Setting a dependency in client is not enough as systemd doesn't know
      when the daemon has initialized itself and started accepting new
      clients. However, it offers a mechanism to solve this. The daemon needs
      to call a special systemd function by which the daemon tells "I'm ready
      to accept new clients". This is exactly what we need with
      libvirtd-guests (client) and libvirtd (daemon). So now, with this
      change, libvirt-guests.service is invoked not any sooner than
      libvirtd.service calls the systemd notify function.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      68954fb2