You need to sign in or sign up before continuing.
  1. 03 8月, 2017 1 次提交
  2. 02 8月, 2017 1 次提交
  3. 21 7月, 2017 1 次提交
    • S
      qemu: Enable NUMA node tag in pci-root for PPC64 · e5a05799
      Shivaprasad G Bhat 提交于
      This patch addresses the same aspects on PPC the bug 1103314 addressed
      on x86.
      
      PCI expander bus creates multiple primary PCI busses, where each of these
      busses can be assigned a specific NUMA affinity, which, on x86 is
      advertised through ACPI on a per-bus basis.
      
      For SPAPR, a PHB's NUMA affinities are assigned on a per-PHB basis, and
      there is no mechanism for advertising NUMA affinities to a guest on a
      per-bus basis. So, even if qemu-ppc manages to get some sort of multi-bus
      topology working using PXB, there is no way to expose the affinities
      of these busses to the guest. It can only be exposed on a per-PHB/per-domain
      basis.
      
      So patch enables NUMA node tag in pci-root controller on PPC.
      
      The way to set the NUMA node is through the numa_node option of
      spapr-pci-host-bridge device. However for the implicit PHB, the only way
      to set the numa_node is from the -global option. The -global option applies
      to all the PHBs unless explicitly specified with the option on the
      respective PHB of CLI. The default PHB has the emulated devices only, so
      the patch prevents setting the NUMA node for the default PHB.
      Signed-off-by: NShivaprasad G Bhat <sbhat@linux.vnet.ibm.com>
      Reviewed-by: NAndrea Bolognani <abologna@redhat.com>
      e5a05799
  4. 20 7月, 2017 1 次提交
  5. 15 7月, 2017 5 次提交
  6. 11 7月, 2017 4 次提交
  7. 20 6月, 2017 1 次提交
  8. 13 6月, 2017 3 次提交
  9. 08 6月, 2017 3 次提交
  10. 26 5月, 2017 1 次提交
  11. 16 5月, 2017 2 次提交
  12. 15 5月, 2017 2 次提交
  13. 21 4月, 2017 1 次提交
    • M
      conf, docs: Add support for coalesce setting(s) · 523c9960
      Martin Kletzander 提交于
      We are currently parsing only rx/frames/max because that's the only
      value that makes sense for us.  The tun device just added support for
      this one and the others are only supported by hardware devices which
      we don't need to worry about as the only way we'd pass those to the
      domain is using <hostdev/> or <interface type='hostdev'/>.  And in
      those cases the guest can modify the settings itself.
      Signed-off-by: NMartin Kletzander <mkletzan@redhat.com>
      523c9960
  14. 11 4月, 2017 2 次提交
  15. 04 4月, 2017 1 次提交
  16. 27 3月, 2017 1 次提交
  17. 17 3月, 2017 2 次提交
  18. 15 3月, 2017 3 次提交
    • M
      qemu: Introduce label-size for NVDIMMs · e433546b
      Michal Privoznik 提交于
      For NVDIMM devices it is optionally possible to specify the size
      of internal storage for namespaces. Namespaces are a feature that
      allows users to partition the NVDIMM for different uses.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      e433546b
    • M
      conf: Introduce @access to <memory/> · 80af11d3
      Michal Privoznik 提交于
      Now that NVDIMM has found its way into libvirt, users might want
      to fine tune some settings for each module separately. One such
      setting is 'share=on|off' for the memory-backend-file object.
      This setting - just like its name suggest already - enables
      sharing the nvdimm module with other applications. Under the hood
      it controls whether qemu mmaps() the file as MAP_PRIVATE or
      MAP_SHARED.
      
      Yet again, we have such config knob in domain XML, but it's just
      an attribute to numa <cell/>. This does not give fine enough
      tuning on per-memdevice basis so we need to have the attribute
      for each device too.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      80af11d3
    • M
      Introduce NVDIMM memory model · b4e8a49f
      Michal Privoznik 提交于
      NVDIMM is new type of memory introduced into QEMU 2.6. The idea
      is that we have a Non-Volatile memory module that keeps the data
      persistent across domain reboots.
      
      At the domain XML level, we already have some representation of
      'dimm' modules. Long story short, NVDIMM will utilize the
      existing <memory/> element that lives under <devices/> by adding
      a new attribute 'nvdimm' to the existing @model and introduce a
      new <path/> element for <source/> while reusing other fields. The
      resulting XML would appear as:
      
          <memory model='nvdimm'>
            <source>
              <path>/tmp/nvdimm</path>
            </source>
            <target>
              <size unit='KiB'>523264</size>
              <node>0</node>
            </target>
            <address type='dimm' slot='0'/>
          </memory>
      
      So far, this is just a XML parser/formatter extension. QEMU
      driver implementation is in the next commit.
      
      For more info on NVDIMM visit the following web page:
      
          http://pmem.io/Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      b4e8a49f
  19. 24 2月, 2017 2 次提交
  20. 09 2月, 2017 1 次提交
    • J
      conf: Add new xml elements for file memorybacking support · bc6d3121
      Jaroslav Safka 提交于
      This part introduces new xml elements for file based
      memorybacking support and their parsing.
      (It allows vhost-user to be used without hugepages.)
      
      New xml elements:
      <memoryBacking>
        <source type="file|anonymous"/>
        <access mode="shared|private"/>
        <allocation mode="immediate|ondemand"/>
      </memoryBacking>
      bc6d3121
  21. 26 1月, 2017 1 次提交
  22. 20 1月, 2017 1 次提交
    • M
      qemu: set default vhost-user ifname · 57b5e27d
      Michal Privoznik 提交于
      Based on work of Mehdi Abaakouk <sileht@sileht.net>.
      
      When parsing vhost-user interface XML and no ifname is found we
      can try to fill it in in post parse callback. The way this works
      is we try to make up interface name from given socket path and
      then ask openvswitch whether it knows the interface.
      Signed-off-by: NMichal Privoznik <mprivozn@redhat.com>
      57b5e27d