1. 18 Feb, 2011 3 commits
  2. 17 Feb, 2011 6 commits
  3. 16 Feb, 2011 2 commits
  4. 15 Feb, 2011 12 commits
  5. 10 Feb, 2011 1 commit
  6. 09 Feb, 2011 3 commits
    • Iustin Pop's avatar
      Fix error msg for instances on offline nodes · 11dcce87
      Iustin Pop authored
      
      
      Currently, for both primary and secondary offline nodes, we give the
      same message:
      - ERROR: instance instance14: instance lives on offline node(s) node3
      - ERROR: instance instance15: instance lives on offline node(s) node3
      - ERROR: instance instance16: instance lives on offline node(s) node3
      - ERROR: instance instance17: instance lives on offline node(s) node3
      
      This is confusing, as an offline primary is in a different category
      than a secondary. The patch changes the warnings to have different
      error messages:
      - ERROR: instance instance14: instance has offline secondary node(s) node3
      - ERROR: instance instance15: instance has offline secondary node(s) node3
      - ERROR: instance instance16: instance lives on offline node node3
      - ERROR: instance instance17: instance lives on offline node node3
      
      Thanks to Alexander Schreiber <als@google.com> for reporting this
      issue.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarAlexander Schreiber <als@google.com>
      11dcce87
    • Iustin Pop's avatar
      cluster verify and instance disks on offline nodes · a3de2ae7
      Iustin Pop authored
      
      
      Currently, cluster-verify says:
      
      - ERROR: instance instance14: couldn't retrieve status for disk/0 on node3: node offline
      - ERROR: instance instance14: instance lives on offline node(s) node3
      - ERROR: instance instance15: couldn't retrieve status for disk/0 on node3: node offline
      - ERROR: instance instance15: instance lives on offline node(s) node3
      
      This is redundant as the “lives on offline node” message should be all we need to
      understand the cluster situation.
      
      The patch fixes this and also corrects a very old idiom.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarStephen Shirley <diamond@google.com>
      a3de2ae7
    • Iustin Pop's avatar
      Cluster verify and N+1 warnings for offline nodes · f7661f6b
      Iustin Pop authored
      
      
      Currently, cluster verify shows warnings N+1 warnings for offline
      nodes having any redundant instances since the memory data that we
      have for those nodes is zero, so any instance will trigger the
      warning.
      
      As the comment says, we already list secondary instances on offline
      nodes, so that warning is enough, and we skip the N+1 one.
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarStephen Shirley <diamond@google.com>
      f7661f6b
  7. 08 Feb, 2011 1 commit
  8. 04 Feb, 2011 11 commits
  9. 03 Feb, 2011 1 commit
    • Iustin Pop's avatar
      Bump up intra-cluster import connect timeout · 81635b5a
      Iustin Pop authored
      
      
      Currently, the export timeout is 10 times 20 seconds, but the import
      is only 30 seconds. I'm raising this to 60 seconds with two goals in
      mind:
      
      - when debugging manually, this allows for easier synchronisation of
        the processes
      - 60 equals to 3 full 20 second intervals, which I think is better
        than just one an a half
      
      This change shouldn't make a big difference either way (at most, it
      will possibly delay the job in case of failures by half a minute).
      Signed-off-by: default avatarIustin Pop <iustin@google.com>
      Reviewed-by: default avatarMichael Hanselmann <hansmi@google.com>
      81635b5a