1. 17 Feb, 2010 1 commit
  2. 15 Feb, 2010 3 commits
    • Iustin Pop's avatar
      Release all node locks during disk replace · d5cd389c
      Iustin Pop authored
      This patch extends commit 7ea7bcf6
      
       by releasing all node locks in disk
      replace for the early release mode. The rationale behind this is:
      
      - LUCreateInstance already releases all node locks while waiting for
        disk synchronization, and does an instance startup later
      - WaitForSync only runs (for disk template 'drbd') 'lvs' and read
        /proc/drbd on the primary node, which should be (modulo bugs in LVM)
        safe for parallel run
      
      In any case, the worst I could foresee is a node having N lvs commands
      run in parallel on it, while being a primary for disk storage. Based on
      create instance doing this safely, and the fact that burnin with more
      than two instances per node is safe, I think this can be applied.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
      d5cd389c
    • Iustin Pop's avatar
      Unify a few re.compile calls in DRBD · 9122e60a
      Iustin Pop authored
      
      
      These are both cleanups and, in the case of _MassageProcData, switching
      from a weaker RE to a stronger one (we now need cs: in the line,
      previosuly any line starting with \d+: was accepted).
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
      9122e60a
    • Iustin Pop's avatar
      Auto-enable early release for offline old nodes · 9af0fa6a
      Iustin Pop authored
      
      
      In case the old node is offline, we won't be able to talk to it to
      remove the storage, and in most cases the node is powered
      off/unreachable.
      
      In this case, it makes no sense to delay the storage release, so we
      enable automatically early_release mode, gaining parallelism during node
      evacuation.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
      9af0fa6a
  3. 11 Feb, 2010 7 commits
  4. 10 Feb, 2010 3 commits
  5. 09 Feb, 2010 1 commit
    • Iustin Pop's avatar
      Add an early release lock/storage for disk replace · 7ea7bcf6
      Iustin Pop authored
      
      
      This patch adds an early_release parameter in the OpReplaceDisks and
      OpEvacuateNode opcodes, allowing earlier release of storage and more
      importantly of internal Ganeti locks.
      
      The behaviour of the early release is that any locks and storage on all
      secondary nodes are released early. This is valid for change secondary
      (where we remove the storage on the old secondary, and release the locks
      on the old and new secondary) and replace on secondary (where we remove
      the old storage and release the lock on the secondary node.
      
      Using this, on a three node setup:
      
      - instance1 on nodes A:B
      - instance2 on nodes C:B
      
      It is possible to run in parallel a replace-disks -s (on secondary) for
      instances 1 and 2.
      
      Replace on primary will remove the storage, but not the locks, as we use
      the primary node later in the LU to check consistency.
      
      It is debatable whether to also remove the locks on the primary node,
      and thus making replace-disks keep zero locks during the sync. While
      this would allow greatly enhanced parallelism, let's first see how
      removal of secondary locks works.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarGuido Trotter <ultrotter@google.com>
      7ea7bcf6
  6. 08 Feb, 2010 1 commit
  7. 03 Feb, 2010 4 commits
  8. 01 Feb, 2010 2 commits
  9. 29 Jan, 2010 1 commit
  10. 28 Jan, 2010 2 commits
    • Guido Trotter's avatar
      Fix flipping MC flag bug · cea0534a
      Guido Trotter authored
      
      
      Currently unofflining or undraining an already functional master
      candidate node, can cause it to demote itself. In order to avoid that we
      only trigger the self-promotion check if the node is not currently a
      candidate.
      Signed-off-by: default avatarGuido Trotter <ultrotter@google.com>
      Reviewed-by: default avatarIustin Pop <iustin@google.com>
      cea0534a
    • Iustin Pop's avatar
      Add capability to use syslog for logging · 551b6283
      Iustin Pop authored
      
      
      This patch adds a configure-time parameter that will set the defaults
      used by all programs, and command-line parameters in the daemons that
      allow overriding it.
      
      Syslog 'yes' enables syslog in addition to file-based logging, 'only'
      enables syslog and disables file-based syslog.
      
      The log entries will be of the form:
      Jan 27 08:45:04 node2 ganeti-noded[14504]: INFO 172.24.227.5:50850 PUT
      /jobqueue_update HTTP/1.0 200
      Jan 27 08:45:05 node2 ganeti-noded[14505]: INFO 172.24.227.5:50853 PUT
      /lv_list HTTP/1.0 200
      
      and (for a multi-threaded program):
      Jan 27 08:51:48 node1 ganeti-masterd[15491]: (MainThread) INFO
      ganeti-masterd daemon startup
      Jan 27 08:51:49 node1 ganeti-masterd[15491]: (MainThread) INFO
      Inspecting job queue
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
      551b6283
  11. 27 Jan, 2010 3 commits
  12. 25 Jan, 2010 4 commits
  13. 22 Jan, 2010 2 commits
  14. 20 Jan, 2010 1 commit
  15. 15 Jan, 2010 1 commit
  16. 14 Jan, 2010 2 commits
  17. 13 Jan, 2010 2 commits