1. 20 3月, 2014 9 次提交
    • A
      Use new dictGetRandomKeys() API to get samples for eviction. · c641b670
      antirez 提交于
      The eviction quality degradates a bit in my tests, but since the API is
      faster, it allows to raise the number of samples, and overall is a win.
      c641b670
    • A
      struct dictEntry -> dictEntry. · 82b53c65
      antirez 提交于
      82b53c65
    • A
      Added dictGetRandomKeys() to dict.c: mass get random entries. · 5317f5e9
      antirez 提交于
      This new function is useful to get a number of random entries from an
      hash table when we just need to do some sampling without particularly
      good distribution.
      
      It just jumps at a random place of the hash table and returns the first
      N items encountered by scanning linearly.
      
      The main usefulness of this function is to speedup Redis internal
      sampling of the key space, for example for key eviction or expiry.
      5317f5e9
    • A
      LRU eviction pool implementation. · 22c9cfaf
      antirez 提交于
      This is an improvement over the previous eviction algorithm where we use
      an eviction pool that is persistent across evictions of keys, and gets
      populated with the best candidates for evictions found so far.
      
      It allows to approximate LRU eviction at a given number of samples
      better than the previous algorithm used.
      22c9cfaf
    • A
      Fix OBJECT IDLETIME return value converting to seconds. · 6d5790d6
      antirez 提交于
      estimateObjectIdleTime() returns a value in milliseconds now, so we need
      to scale the output of OBJECT IDLETIME to seconds.
      6d5790d6
    • A
      Obtain LRU clock in a resolution dependent way. · ad6b0f70
      antirez 提交于
      For testing purposes it is handy to have a very high resolution of the
      LRU clock, so that it is possible to experiment with scripts running in
      just a few seconds how the eviction algorithms works.
      
      This commit allows Redis to use the cached LRU clock, or a value
      computed on demand, depending on the resolution. So normally we have the
      good performance of a precomputed value, and a clock that wraps in many
      days using the normal resolution, but if needed, changing a define will
      switch behavior to an high resolution LRU clock.
      ad6b0f70
    • A
      Specify lruclock in redisServer structure via REDIS_LRU_BITS. · 1faf8266
      antirez 提交于
      The padding field was totally useless: removed.
      1faf8266
    • A
      Specify LRU resolution in milliseconds. · d77e2316
      antirez 提交于
      d77e2316
    • A
      Set LRU parameters via REDIS_LRU_BITS define. · fe308470
      antirez 提交于
      fe308470
  2. 19 3月, 2014 1 次提交
  3. 18 3月, 2014 4 次提交
    • A
      Sentinel: sentinelRefreshInstanceInfo() minor refactoring. · b9e90a70
      antirez 提交于
      Test sentinel.tilt condition on top and return if it is true.
      This allows to remove the check for the tilt condition in the remaining
      code paths of the function.
      b9e90a70
    • A
      218cc5fc
    • A
      Sentinel: down-after-milliseconds is not master-specific. · bb6d8501
      antirez 提交于
      addReplySentinelRedisInstance() modified so that this field is displayed
      for all the kind of instances: Sentinels, Masters, Slaves.
      bb6d8501
    • A
      Sentinel failure detection implementation improved. · ae0b7680
      antirez 提交于
      Failure detection in Sentinel is ping-pong based. It used to work by
      remembering the last time a valid PONG reply was received, and checking
      if the reception time was too old compared to the current current time.
      
      PINGs were sent at a fixed interval of 1 second.
      
      This works in a decent way, but does not scale well when we want to set
      very small values of "down-after-milliseconds" (this is the node
      timeout basically).
      
      This commit reiplements the failure detection making a number of
      changes. Some changes are inspired to Redis Cluster failure detection
      code:
      
      * A new last_ping_time field is added in representation of instances.
        If non zero, we have an active ping that was sent at the specified
        time. When a valid reply to ping is received, the field is zeroed
        again.
      * last_ping_time is not reset when we reconnect the link or send a new
        ping, so from our point of view it represents the time we started
        waiting for the instance to reply to our pings without receiving a
        reply.
      * last_ping_time is now used in order to check if the instance is
        timed out. This means that we can have a node timeout of 100
        milliseconds and yet the system will work well since the new check is
        not bound to the period used to send pings.
      * Pings are now sent every second, or often if the value of
        down-after-milliseconds is less than one second. With a lower limit of
        10 HZ ping frequency.
      * Link reconnection code was improved. This is used in order to try to
        reconnect the link when we are at 50% of the node timeout without a
        valid reply received yet. However the old code triggered unnecessary
        reconnections when the node timeout was very small. Now that should be
        ok.
      
      The new code passes the tests but more testing is needed and more unit
      tests stressing the failure detector, so currently this is merged only
      in the unstable branch.
      ae0b7680
  4. 15 3月, 2014 2 次提交
  5. 14 3月, 2014 3 次提交
    • A
      Sentinel: be safe under crash-recovery assumptions. · ed813863
      antirez 提交于
      Sentinel's main safety argument is that there are no two configurations
      for the same master with the same version (configuration epoch).
      
      For this to be true Sentinels require to be authorized by a majority.
      Additionally Sentinels require to do two important things:
      
      * Never vote again for the same epoch.
      * Never exchange an old vote for a fresh one.
      
      The first prerequisite, in a crash-recovery system model, requires to
      persist the master->leader_epoch on durable storage before to reply to
      messages. This was not the case.
      
      We also make sure to persist the current epoch in order to never reply
      to stale votes requests from other Sentinels, after a recovery.
      
      The configuration is persisted by making use of fsync(), this is
      considered in the context of this code a good enough guarantee that
      after a restart our durable state is restored, however this may not
      always be the case depending on the kind of hardware and operating
      system used.
      ed813863
    • A
      Sentinel: fake PUBLISH command to receive HELLO messages. · 36509402
      antirez 提交于
      Now the way HELLO messages are received is unified.
      Now it is no longer needed for Sentinels to converge to the higher
      configuration for a master to be able to chat via some Redis instance,
      the are able to directly exchanges configurations.
      
      Note that this commit does not include the (trivial) change needed to
      send HELLO messages to Sentinel instances as well, since for an error I
      committed the change in the previous commit that refactored hello
      messages processing into a separated function.
      36509402
    • A
  6. 13 3月, 2014 2 次提交
  7. 11 3月, 2014 9 次提交
  8. 10 3月, 2014 10 次提交