1. 04 Aug, 2011 2 commits
  2. 03 Aug, 2011 1 commit
  3. 02 Aug, 2011 1 commit
  4. 26 Jul, 2011 2 commits
  5. 25 Jul, 2011 1 commit
  6. 22 Jul, 2011 2 commits
  7. 21 Jul, 2011 3 commits
  8. 12 Jul, 2011 1 commit
  9. 08 Jul, 2011 1 commit
  10. 05 Jul, 2011 2 commits
  11. 15 Jun, 2011 1 commit
  12. 07 Jun, 2011 1 commit
  13. 01 Jun, 2011 1 commit
  14. 31 May, 2011 2 commits
  15. 30 May, 2011 1 commit
    • Michael Hanselmann's avatar
      gnt-node migrate: Use LU-generated jobs · b7a1c816
      Michael Hanselmann authored
      Until now LUNodeMigrate used multiple tasklets to evacuate all primary
      instances on a node. In some cases it would acquire all node locks,
      which isn't good on big clusters. With upcoming improvements to the LUs
      for instance failover and migration, switching to separate jobs looks
      like a better option. This patch changes LUNodeMigrate to use
      LU-generated jobs.
      While working on this patch, I identified a race condition in
      LUNodeMigrate.ExpandNames. A node's instances were retrieved without a
      lock and no verification was done.
      For RAPI, a new feature string is added and can be used to detect
      clusters which support more parameters for node migration. The client
      is updated.
      Signed-off-by: default avatarMichael Hanselmann <hansmi@google.com>
      Reviewed-by: default avatarIustin Pop <iustin@google.com>
      Signed-off-by: default avatarMichael Hanselmann <hansmi@google.com>
  16. 25 May, 2011 1 commit
    • Michael Hanselmann's avatar
      Fix bug in LUNodeMigrate · f8fa4175
      Michael Hanselmann authored
      Commit aac4511a
       added CheckArguments to LUNodeMigrate with a call to
      _CheckIAllocatorOrNode. When no default iallocator is defined,
      evacuating a node would always fail:
      $ gnt-node migrate node123
      Migrate instance(s) '...'?
      y/[n]/?: y
      Failure: prerequisites not met for this operation:
      No iallocator or node given and no cluster-wide default iallocator
      found; please specify either an iallocator or a node, or set a
      cluster-wide default iallocator
      This patch adds a new parameter to specify a target node. This doesn't
      solve all issues, but will make the most important cases work again in
      the meantime. This opcode will receive more work for node group support.
      Signed-off-by: default avatarMichael Hanselmann <hansmi@google.com>
      Reviewed-by: default avatarIustin Pop <iustin@google.com>
  17. 24 May, 2011 1 commit
  18. 20 May, 2011 3 commits
  19. 16 May, 2011 2 commits
  20. 12 May, 2011 3 commits
  21. 09 May, 2011 2 commits
    • Iustin Pop's avatar
      Add --no-wait-for-sync when converting to drbd · 456798ab
      Iustin Pop authored
      Currently, when converting an instance from plain to DRBD, the
      instance is blocked during the entire resync period. This patch adds
      the --no-wait-for-sync so that the operation finishes as soon as the
      DRBD sync has started, without waiting for the entire sync. This makes
      the instance available much faster.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
    • Iustin Pop's avatar
      Recreate instance disks: allow changing nodes · c8a96ae7
      Iustin Pop authored
      This patch introduces the option of changing an instance's nodes when
      doing the disk recreation. The rationale is that currently if an
      instance lives on a node that has gone down and is marked offline,
      it's not possible to re-create the disks and reinstall the instance on
      a different node without hacking the config file.
      Additionally, the LU now locks the instance's nodes (which was not
      done before), as we most likely allocate new resources on them.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarRené Nussbaumer <rn@google.com>
  22. 29 Apr, 2011 1 commit
  23. 27 Apr, 2011 1 commit
    • Iustin Pop's avatar
      Prevent readding of the master node · d833acc6
      Iustin Pop authored
      This breaks Ganeti in multiple ways. If we don't make the check in
      gnt-node itself, then bootstrap.SetupNodeDaemon will restart the
      master daemon, making the operation fail:
        node1# gnt-node add --readd node1
        Cannot communicate with the master daemon.
        Is it running and listening for connections?
      The check in cmdlib is more of a safety check, as we shouldn't reach
      it. If we do (via a bad client), then it will prevent breakage in the
      job queue/config handling.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
  24. 21 Apr, 2011 1 commit
  25. 20 Apr, 2011 1 commit
  26. 19 Apr, 2011 2 commits