Skip to content
Snippets Groups Projects
  1. Oct 25, 2010
  2. Oct 21, 2010
  3. Oct 07, 2010
  4. Oct 06, 2010
  5. Sep 15, 2010
    • Iustin Pop's avatar
      hbal: implement user-friendly termination requests · 03cb89f0
      Iustin Pop authored
      Currently, hbal will abort immediately when requested (^C, or SIGINT,
      etc.). This is not nice, since then the already started jobs need to be
      tracked manually.
      
      This patch adds a signal handler for SIGINT and SIGTERM, which will, the
      first time, simply record the shutdown request (and hbal will then exit
      once all jobs in the current jobset finish), and at the second request,
      will cause an immediate exit.
      03cb89f0
  6. Sep 03, 2010
  7. Sep 02, 2010
    • Iustin Pop's avatar
      Makefile: make the rst2html converter more strict · d78ceb9e
      Iustin Pop authored
      This will make the automated builds flag any problems.
      d78ceb9e
    • Iustin Pop's avatar
      Add some more debugging functions · adc5c176
      Iustin Pop authored
      These are just variations of the standard debug, but are provided for
      simpler code, since lazyness is something causing non-computation of
      debug statements.
      adc5c176
    • Iustin Pop's avatar
      Fix ReplaceSecondary moves for offline nodes · 74e89a14
      Iustin Pop authored
      The addition of a new secondary on a node is doing two memory tests:
      - in strict mode, reject if we get into N+1 failure
      - reject if the new instance memory is greater than the free memory (not
        available memory) on the node
      
      The last check is designed to ensure that, irrespective of the other
      secondary instances on this node, we are able to failover/migrate the
      newly-added instance.
      
      However, we should allow this, if the instances comes from an offline
      node, which doesn't offer anything (not even disk replication).
      Therefore this patch makes this check conditional on the strict mode.
      74e89a14
    • Iustin Pop's avatar
      Update NEWS file · 49d977db
      Iustin Pop authored
      49d977db
  8. Aug 30, 2010
  9. Aug 25, 2010
    • Iustin Pop's avatar
      Add a new option --save-cluster · 02da9d07
      Iustin Pop authored
      This option will in the future be used to serialize the cluster state in
      hbal and hspace after the rebalance/allocation steps.
      02da9d07
    • Iustin Pop's avatar
      Add unittest for Node text serialization · 50811e2c
      Iustin Pop authored
      This checks that the Node text serialization and deserialization
      operations are idempotent when combined other.
      50811e2c
    • Iustin Pop's avatar
      Switch unittest to custom hostnames · a070c426
      Iustin Pop authored
      Currently, the hostnames are almost fully arbitrary chars, which breaks
      the assumption that nodes/instances will be normal DNS hostnames.
      
      This patch adds some custom generators for these hostnames, that will
      allow better testing of text loader serialization/deserialization.
      a070c426
  10. Aug 24, 2010
  11. Jul 29, 2010
  12. Jul 27, 2010
  13. Jul 23, 2010
  14. Jul 22, 2010
  15. Jul 21, 2010
    • Iustin Pop's avatar
      Use --union for hpc sum · 7e9e8245
      Iustin Pop authored
      … which fixes the issue noted in the previous commit (almost a brown
      paper bag change).
      7e9e8245
    • Iustin Pop's avatar
      Preliminary support for coverage during live-test · dc61c50b
      Iustin Pop authored
      While this doesn't work correctly yet (hpc sum seems to only take common
      modules, not the sum of modules?), it prepares for gathering coverage
      data during live-test (as an alternative to unittest coverage data).
      dc61c50b
    • Iustin Pop's avatar
      Add some more imports to QC.hs · 223dbe53
      Iustin Pop authored
      This is needed so that in the coverage report we list all modules, even
      the ones we don't test at all, such that we get the complete results.
      223dbe53
    • Iustin Pop's avatar
      Change the meaning of the N+1 fail metric · c3c7a0c1
      Iustin Pop authored
      Currently, this metric tracks the nodes failing the N+1 check. While
      this helps (in some cases) to evacuate such nodes, it's not a good
      metric since rarely it will change during a step (only at the last
      instance moving away). Therefore we replace it with the count of
      instances living on such nodes, which is much better because:
      - moving an instance away while the node is still N+1 failing will still
        reflect in the score as an optimization
      - moving the last instance causing an N+1 failure will result in a heavy
        decrease of this score, thus giving the right bonus to clear this
        status
      c3c7a0c1
    • Iustin Pop's avatar
      Introduce per-metric weights · 8a3b30ca
      Iustin Pop authored
      Currently all metrics have the same weight (we just sum them together).
      However, for the hard constraints (N+1 failures, offline nodes, etc.)
      we should handle the metrics differently based on their meaning. For
      example, an instance living on a primary offline node is worse than an
      instance having its secondary node offline, which in turn is worse than
      an instance having its secondary node failing N+1.
      
      To express this case in our code, we introduce a table of weights for
      the metrics, with which we can influence their relative importance.
      8a3b30ca
    • Iustin Pop's avatar
      Allow balancing moves to introduce N+1 errors · 2cae47e9
      Iustin Pop authored
      This patch switches the applyMove function to the extended versions of
      Node.addPri and addSec, and passes the override flag based on the state
      of the node that we're moving away from.
      2cae47e9
    • Iustin Pop's avatar
      Introduce a relaxed add instance mode · 3e3c9393
      Iustin Pop authored
      In case an instance is living on an offline node, it doesn't make sense
      to refuse moving it because that would create N+1 failures; failing N+1
      is still much better than not running at all. Similarly, if the
      secondary node of an instance is offline, meaning the instance doesn't
      have any redundancy, we have a worse case than having a secondary that
      is N+1 failing and it could not accept the instance as primary, but it
      stil does redundancy for it.
      
      To allow this, we rename Node.addPri to addPriEx and introduce an extra
      parameter (addPri is a partial application of addPriEx and keeps the
      same signature). Node.addSec gets the same treatement.
      3e3c9393
Loading