“bd3569be1b6bcc33885dc8c7413e20f5e6fd5315”上不存在“fleetrec/models/recall/word2vec/sample_test.txt”
  1. 26 9月, 2013 1 次提交
    • L
      qemu: Adjust qemu wakeup · 4bc78a87
      Liu, Jinsong 提交于
      Currently Xen hvm s3 has a bug coming from the difference between
      qemu-traditioanl and qemu-xen. For qemu-traditional, the way to
      resume from hvm s3 is via 'xl trigger' command. However, for
      qemu-xen, the way to resume from hvm s3 inherited from standard
      qemu, i.e. via QMP, and it doesn't work under Xen.
      
      The root cause is, for qemu-xen, 'xl trigger' command didn't reset
      devices, while QMP didn't unpause hvm domain though they did qemu
      system reset.
      
      We have two qemu patches and one xl patch to fix Xen hvm s3 bug.
      This patch is the qemu patch 1. It adjusts qemu wakeup so that
      Xen s3 resume logic (which will be implemented at qemu patch 2)
      will be notified after qemu system reset.
      Signed-off-by: NLiu Jinsong <jinsong.liu@intel.com>
      Signed-off-by: NStefano Stabellini <stefano.stabellini@eu.citrix.com>
      Reviewed-by: NPaolo Bonzini <pbonzini@redhat.com>
      Reviewed-by: NAnthony PERARD <anthony.perard@citrix.com>
      4bc78a87
  2. 12 9月, 2013 1 次提交
  3. 23 8月, 2013 1 次提交
  4. 30 7月, 2013 1 次提交
  5. 15 7月, 2013 1 次提交
  6. 10 7月, 2013 1 次提交
  7. 01 5月, 2013 2 次提交
  8. 30 4月, 2013 1 次提交
  9. 27 4月, 2013 1 次提交
  10. 11 3月, 2013 1 次提交
  11. 22 2月, 2013 1 次提交
    • A
      ui: add basic GTK gui (v5) · a4ccabcf
      Anthony Liguori 提交于
      This is minimalistic and just contains the basic widget infrastructure.  The GUI
      consists of a menu and a GtkNotebook.  To start with, the notebook has its tabs
      hidden which provides a UI that looks very similar to SDL with the exception of
      the menu bar.
      
      The menu bar allows a user to toggle the visibility of the tabs.  Cairo is used
      for rendering.
      
      I used gtk-vnc as a reference.  gtk-vnc solves the same basic problems as QEMU
      since it was originally written as a remote display for QEMU.  So for the most
      part, the approach to rendering and keyboard handling should be pretty solid for
      GTK.
      Signed-off-by: NAnthony Liguori <aliguori@us.ibm.com>
      Message-id: 1361367806-4599-4-git-send-email-aliguori@us.ibm.com
      a4ccabcf
  12. 21 2月, 2013 1 次提交
  13. 26 1月, 2013 1 次提交
  14. 19 1月, 2013 1 次提交
  15. 17 1月, 2013 2 次提交
  16. 13 1月, 2013 1 次提交
  17. 21 12月, 2012 1 次提交
    • J
      savevm: New save live migration method: pending · e4ed1541
      Juan Quintela 提交于
      Code just now does (simplified for clarity)
      
          if (qemu_savevm_state_iterate(s->file) == 1) {
             vm_stop_force_state(RUN_STATE_FINISH_MIGRATE);
             qemu_savevm_state_complete(s->file);
          }
      
      Problem here is that qemu_savevm_state_iterate() returns 1 when it
      knows that remaining memory to sent takes less than max downtime.
      
      But this means that we could end spending 2x max_downtime, one
      downtime in qemu_savevm_iterate, and the other in
      qemu_savevm_state_complete.
      
      Changed code to:
      
          pending_size = qemu_savevm_state_pending(s->file, max_size);
          DPRINTF("pending size %lu max %lu\n", pending_size, max_size);
          if (pending_size >= max_size) {
              ret = qemu_savevm_state_iterate(s->file);
           } else {
              vm_stop_force_state(RUN_STATE_FINISH_MIGRATE);
              qemu_savevm_state_complete(s->file);
           }
      
      So what we do is: at current network speed, we calculate the maximum
      number of bytes we can sent: max_size.
      
      Then we ask every save_live section how much they have pending.  If
      they are less than max_size, we move to complete phase, otherwise we
      do an iterate one.
      
      This makes things much simpler, because now individual sections don't
      have to caluclate the bandwidth (it was implossible to do right from
      there).
      Signed-off-by: NJuan Quintela <quintela@redhat.com>
      Reviewed-by: NPaolo Bonzini <pbonzini@redhat.com>
      e4ed1541
  18. 19 12月, 2012 2 次提交
  19. 07 12月, 2012 1 次提交
  20. 06 12月, 2012 1 次提交
  21. 02 11月, 2012 1 次提交
  22. 29 10月, 2012 1 次提交
    • Z
      Add USB option in machine options · 094b287f
      zhlcindy@gmail.com 提交于
      When -usb option is used, global varible usb_enabled is set.
      And all the plaform will create one USB controller according
      to this variable. In fact, global varibles make code hard
      to read.
      
      So this patch is to remove global variable usb_enabled and
      add USB option in machine options. All the plaforms will get
      USB option value from machine options.
      
      USB option of machine options will be set either by:
        * -usb
        * -machine type=pseries,usb=on
      
      Both these ways can work now. They both set USB option in
      machine options. In the future, the first way will be removed.
      Signed-off-by: NLi Zhang <zhlcindy@linux.vnet.ibm.com>
      Acked-by: NAlexander Graf <agraf@suse.de>
      Signed-off-by: NAlexander Graf <agraf@suse.de>
      094b287f
  23. 07 10月, 2012 1 次提交
  24. 26 9月, 2012 2 次提交
  25. 27 8月, 2012 1 次提交
  26. 24 8月, 2012 1 次提交
  27. 17 8月, 2012 1 次提交
    • D
      Allow QEMUMachine to override reset sequencing · be522029
      David Gibson 提交于
      qemu_system_reset() function always performs the same basic actions on
      all machines.  This includes running all the reset handler hooks,
      however the order in which these will run is not always easily predictable.
      
      This patch splits the core of qemu_system_reset() - the invocation of
      the reset handlers - out into a new qemu_devices_reset() function.
      qemu_system_reset() will usually call qemu_devices_reset(), but that
      can be now overriden by a new reset method in the QEMUMachine
      structure.
      
      Individual machines can use this reset method, if necessary, to
      perform any extra, machine specific initializations which have to
      occur before or after the bulk of the reset handlers.  It's expected
      that the method will call qemu_devices_reset() at some point, but if
      the machine has really strange ordering requirements between devices
      resets it could even override that with it's own reset sequence (with
      great care, obviously).
      
      For a specific example of when this might be needed: a number of
      machines (but not PC) load images specified with -kernel or -initrd
      directly into the machine RAM before booting the guest.  This mostly
      works at the moment, but to make this actually safe requires that this
      load occurs after peripheral devices are reset - otherwise they could
      have active DMAs in progress which would clobber the in memory images.
      Some machines (notably pseries) also have other entry conditions which
      need to be set up as the last thing before executing in guest space -
      some of this could be considered "emulated firmware" in the sense that
      the actions of the firmware are emulated directly by qemu rather than
      by executing a firmware image within the guest.  When the platform's
      firmware to OS interface is sufficiently well specified, this saves
      time both in implementing the "firmware" and executing it.
      
      aliguori: don't unconditionally dereference current_machine
      Reviewed-by: NAndreas Färber <afaerber@suse.de>
      Signed-off-by: NDavid Gibson <david@gibson.dropbear.id.au>
      Signed-off-by: NAnthony Liguori <aliguori@us.ibm.com>
      be522029
  28. 04 8月, 2012 1 次提交
    • C
      Fixes related to processing of qemu's -numa option · ee785fed
      Chegu Vinod 提交于
      The -numa option to qemu is used to create [fake] numa nodes
      and expose them to the guest OS instance.
      
      There are a couple of issues with the -numa option:
      
      a) Max VCPU's that can be specified for a guest while using
         the qemu's -numa option is 64. Due to a typecasting issue
         when the number of VCPUs is > 32 the VCPUs don't show up
         under the specified [fake] numa nodes.
      
      b) KVM currently has support for 160VCPUs per guest. The
         qemu's -numa option has only support for upto 64VCPUs
         per guest.
      This patch addresses these two issues.
      
      Below are examples of (a) and (b)
      
      a) >32 VCPUs are specified with the -numa option:
      
      /usr/local/bin/qemu-system-x86_64 \
      -enable-kvm \
      71:01:01 \
      -net tap,ifname=tap0,script=no,downscript=no \
      -vnc :4
      
      ...
      Upstream qemu :
      --------------
      
      QEMU 1.1.50 monitor - type 'help' for more information
      (qemu) info numa
      6 nodes
      node 0 cpus: 0 1 2 3 4 5 6 7 8 9 32 33 34 35 36 37 38 39 40 41
      node 0 size: 131072 MB
      node 1 cpus: 10 11 12 13 14 15 16 17 18 19 42 43 44 45 46 47 48 49 50 51
      node 1 size: 131072 MB
      node 2 cpus: 20 21 22 23 24 25 26 27 28 29 52 53 54 55 56 57 58 59
      node 2 size: 131072 MB
      node 3 cpus: 30
      node 3 size: 131072 MB
      node 4 cpus:
      node 4 size: 131072 MB
      node 5 cpus: 31
      node 5 size: 131072 MB
      
      With the patch applied :
      -----------------------
      
      QEMU 1.1.50 monitor - type 'help' for more information
      (qemu) info numa
      6 nodes
      node 0 cpus: 0 1 2 3 4 5 6 7 8 9
      node 0 size: 131072 MB
      node 1 cpus: 10 11 12 13 14 15 16 17 18 19
      node 1 size: 131072 MB
      node 2 cpus: 20 21 22 23 24 25 26 27 28 29
      node 2 size: 131072 MB
      node 3 cpus: 30 31 32 33 34 35 36 37 38 39
      node 3 size: 131072 MB
      node 4 cpus: 40 41 42 43 44 45 46 47 48 49
      node 4 size: 131072 MB
      node 5 cpus: 50 51 52 53 54 55 56 57 58 59
      node 5 size: 131072 MB
      
      b) >64 VCPUs specified with -numa option:
      
      /usr/local/bin/qemu-system-x86_64 \
      -enable-kvm \
      -cpu Westmere,+rdtscp,+pdpe1gb,+dca,+pdcm,+xtpr,+tm2,+est,+smx,+vmx,+ds_cpl,+monitor,+dtes64,+pclmuldq,+pbe,+tm,+ht,+ss,+acpi,+d-vnc :4
      
      ...
      
      Upstream qemu :
      --------------
      
      only 63 CPUs in NUMA mode supported.
      only 64 CPUs in NUMA mode supported.
      QEMU 1.1.50 monitor - type 'help' for more information
      (qemu) info numa
      8 nodes
      node 0 cpus: 6 7 8 9 38 39 40 41 70 71 72 73
      node 0 size: 65536 MB
      node 1 cpus: 10 11 12 13 14 15 16 17 18 19 42 43 44 45 46 47 48 49 50 51 74 75 76 77 78 79
      node 1 size: 65536 MB
      node 2 cpus: 20 21 22 23 24 25 26 27 28 29 52 53 54 55 56 57 58 59 60 61
      node 2 size: 65536 MB
      node 3 cpus: 30 62
      node 3 size: 65536 MB
      node 4 cpus:
      node 4 size: 65536 MB
      node 5 cpus:
      node 5 size: 65536 MB
      node 6 cpus: 31 63
      node 6 size: 65536 MB
      node 7 cpus: 0 1 2 3 4 5 32 33 34 35 36 37 64 65 66 67 68 69
      node 7 size: 65536 MB
      
      With the patch applied :
      -----------------------
      
      QEMU 1.1.50 monitor - type 'help' for more information
      (qemu) info numa
      8 nodes
      node 0 cpus: 0 1 2 3 4 5 6 7 8 9
      node 0 size: 65536 MB
      node 1 cpus: 10 11 12 13 14 15 16 17 18 19
      node 1 size: 65536 MB
      node 2 cpus: 20 21 22 23 24 25 26 27 28 29
      node 2 size: 65536 MB
      node 3 cpus: 30 31 32 33 34 35 36 37 38 39
      node 3 size: 65536 MB
      node 4 cpus: 40 41 42 43 44 45 46 47 48 49
      node 4 size: 65536 MB
      node 5 cpus: 50 51 52 53 54 55 56 57 58 59
      node 5 size: 65536 MB
      node 6 cpus: 60 61 62 63 64 65 66 67 68 69
      node 6 size: 65536 MB
      node 7 cpus: 70 71 72 73 74 75 76 77 78 79
      Signed-off-by: NChegu Vinod &lt;chegu_vinod@hp.com&gt;, Jim Hull &lt;jim.hull@hp.com&gt;, Craig Hada <craig.hada@hp.com>
      Tested-by: NEduardo Habkost <ehabkost@redhat.com>
      Reviewed-by: NEduardo Habkost <ehabkost@redhat.com>
      Signed-off-by: NBlue Swirl <blauwirbel@gmail.com>
      ee785fed
  29. 29 6月, 2012 1 次提交
  30. 15 3月, 2012 2 次提交
    • L
      qapi: Convert migrate · e1c37d0e
      Luiz Capitulino 提交于
      The migrate command is one of those commands where HMP and QMP completely
      mix up together. This made the conversion to the QAPI (which separates the
      command into QMP and HMP parts) a bit difficult.
      
      The first important change to be noticed is that this commit completes the
      removal of the Monitor object from migration code, started by the previous
      commit.
      
      Another important and tricky change is about supporting the non-detached
      mode. That is, if the user doesn't pass '-d' the migrate command will lock
      the monitor and will only release it when migration is finished.
      
      To support this in the new HMP command (hmp_migrate()), it is necessary
      to create a timer which runs every second and checks if the migration is
      still active. If it is, the timer callback will re-schedule itself to run
      one second in the future. If the migration has already finished, the
      monitor lock is released and the user can use it normally.
      
      All these changes should be transparent to the user.
      Signed-off-by: NAnthony Liguori <aliguori@us.ibm.com>
      Signed-off-by: NLuiz Capitulino <lcapitulino@redhat.com>
      e1c37d0e
    • L
      Purge migration of (almost) everything to do with monitors · 539de124
      Luiz Capitulino 提交于
      The Monitor object is passed back and forth within the migration/savevm
      code so that it can print errors and progress to the user.
      
      However, that approach assumes a HMP monitor, being completely invalid
      in QMP.
      
      This commit drops almost every single usage of the Monitor object, all
      monitor_printf() calls have been converted into DPRINTF() ones.
      
      There are a few remaining Monitor objects, those are going to be dropped
      by the next commit.
      Signed-off-by: NAnthony Liguori <aliguori@us.ibm.com>
      Signed-off-by: NLuiz Capitulino <lcapitulino@redhat.com>
      539de124
  31. 25 2月, 2012 3 次提交
  32. 02 2月, 2012 1 次提交
  33. 19 1月, 2012 1 次提交