1. 18 4月, 2014 1 次提交
  2. 16 4月, 2014 13 次提交
  3. 25 3月, 2014 6 次提交
    • A
      adjustOpenFilesLimit() refactoring. · 3580bb48
      antirez 提交于
      In this commit:
      * Decrement steps are semantically differentiated from the reserved FDs.
        Previously both values were 32 but the meaning was different.
      * Make it clear that we save setrlimit errno.
      * Don't explicitly handle wrapping of 'f', but prevent it from
        happening.
      * Add comments to make the function flow more readable.
      
      This integrates PR #1630
      3580bb48
    • M
      Fix potentially incorrect errno usage · c3510af1
      Matt Stancliff 提交于
      errno may be reset by the previous call to redisLog, so capture
      the original value for proper error reporting.
      c3510af1
    • M
      Add REDIS_MIN_RESERVED_FDS define for open fds · 771f8ad0
      Matt Stancliff 提交于
      Also update the original REDIS_EVENTLOOP_FDSET_INCR to
      include REDIS_MIN_RESERVED_FDS. REDIS_EVENTLOOP_FDSET_INCR
      exists to make sure more than (maxclients+RESERVED) entries
      are allocated, but we can only guarantee that if we include
      the current value of REDIS_MIN_RESERVED_FDS as a minimum
      for the INCR size.
      771f8ad0
    • M
      Fix infinite loop on startup if ulimit too low · 3ce742d1
      Matt Stancliff 提交于
      Fun fact: rlim_t is an unsigned long long on all platforms.
      
      Continually subtracting from a rlim_t makes it get smaller
      and smaller until it wraps, then you're up to 2^64-1.
      
      This was causing an infinite loop on Redis startup if
      your ulimit was extremely (almost comically) low.
      
      The case of (f > oldlimit) would never be met in a case like:
      
          f = 150
          while (f > 20) f -= 128
      
      Since f is unsigned, it can't go negative and would
      take on values of:
      
          Iteration 1: 150 - 128 => 22
          Iteration 2:  22 - 128 => 18446744073709551510
          Iterations 3-∞: ...
      
      To catch the wraparound, we use the previous value of f
      stored in limit.rlimit_cur.  If we subtract from f and
      get a larger number than the value it had previously,
      we print an error and exit since we don't have enough
      file descriptors to help the user at this point.
      
      Thanks to @bs3g for the inspiration to fix this problem.
      Patches existed from @bs3g at antirez#1227, but I needed to repair a few other
      parts of Redis simultaneously, so I didn't get a chance to use them.
      3ce742d1
    • M
      Improve error handling around setting ulimits · b5958539
      Matt Stancliff 提交于
      The log messages about open file limits have always
      been slightly opaque and confusing.  Here's an attempt to
      fix their wording, detail, and meaning.  Users will have a
      better understanding of how to fix very common problems
      with these reworded messages.
      
      Also, we handle a new error case when maxclients becomes less
      than one, essentially rendering the server unusable.  We
      now exit on startup instead of leaving the user with a server
      unable to handle any connections.
      
      This fixes antirez#356 as well.
      b5958539
    • M
      Replace magic 32 with REDIS_EVENTLOOP_FDSET_INCR · 6826af1b
      Matt Stancliff 提交于
      32 was the additional number of file descriptors Redis
      would reserve when managing a too-low ulimit.  The
      number 32 was in too many places statically, so now
      we use a macro instead that looks more appropriate.
      
      When Redis sets up the server event loop, it uses:
          server.maxclients+REDIS_EVENTLOOP_FDSET_INCR
      
      So, when reserving file descriptors, it makes sense to
      reserve at least REDIS_EVENTLOOP_FDSET_INCR FDs instead
      of only 32.  Currently, REDIS_EVENTLOOP_FDSET_INCR is
      set to 128 in redis.h.
      
      Also, I replaced the static 128 in the while f < old loop
      with REDIS_EVENTLOOP_FDSET_INCR as well, which results
      in no change since it was already 128.
      
      Impact: Users now need at least maxclients+128 as
      their open file limit instead of maxclients+32 to obtain
      actual "maxclients" number of clients.  Redis will carve
      the extra REDIS_EVENTLOOP_FDSET_INCR file descriptors it
      needs out of the "maxclients" range instead of failing
      to start (unless the local ulimit -n is too low to accomidate
      the request).
      6826af1b
  4. 24 3月, 2014 2 次提交
    • A
      Sample and cache RSS in serverCron(). · 4ebc7e37
      antirez 提交于
      Obtaining the RSS (Resident Set Size) info is slow in Linux and OSX.
      This slowed down the generation of the INFO 'memory' section.
      
      Since the RSS does not require to be a real-time measurement, we
      now sample it with server.hz frequency (10 times per second by default)
      and use this value both to show the INFO rss field and to compute the
      fragmentation ratio.
      
      Practically this does not make any difference for memory profiling of
      Redis but speeds up the INFO call significantly.
      4ebc7e37
    • A
      Cache uname() output across INFO calls. · 7054f2a0
      antirez 提交于
      Uname was profiled to be a slow syscall. It produces always the same
      output in the context of a single execution of Redis, so calling it at
      every INFO output generation does not make too much sense.
      
      The uname utsname structure was modified as a static variable. At the
      same time a static integer was added to check if we need to call uname
      the first time.
      7054f2a0
  5. 21 3月, 2014 1 次提交
  6. 05 3月, 2014 2 次提交
    • A
      Document why we update peak memory in INFO. · 55b8f6ec
      antirez 提交于
      55b8f6ec
    • M
      Force INFO used_memory_peak to match peak memory · 647a2614
      Matt Stancliff 提交于
      used_memory_peak only updates in serverCron every server.hz,
      but Redis can use more memory and a user can request memory
      INFO before used_memory_peak gets updated in the next
      cron run.
      
      This patch updates used_memory_peak to the current
      memory usage if the current memory usage is higher
      than the recorded used_memory_peak value.
      
      (And it only calls zmalloc_used_memory() once instead of
      twice as it was doing before.)
      647a2614
  7. 27 2月, 2014 2 次提交
    • A
      Initial implementation of BITPOS. · 3294f74f
      antirez 提交于
      It appears to work but more stress testing, and both unit tests and
      fuzzy testing, is needed in order to ensure the implementation is sane.
      3294f74f
    • A
      Initial implementation of BITPOS. · 25e2791e
      antirez 提交于
      It appears to work but more stress testing, and both unit tests and
      fuzzy testing, is needed in order to ensure the implementation is sane.
      25e2791e
  8. 17 2月, 2014 1 次提交
    • A
      Get absoulte config file path before processig 'dir'. · 4237f14a
      antirez 提交于
      The code tried to obtain the configuration file absolute path after
      processing the configuration file. However if config file was a relative
      path and a "dir" statement was processed reading the config, the absolute
      path obtained was wrong.
      
      With this fix the absolute path is obtained before processing the
      configuration while the server is still in the original directory where
      it was executed.
      4237f14a
  9. 13 2月, 2014 1 次提交
    • A
      Update cached time in rdbLoad() callback. · 85492dcf
      antirez 提交于
      server.unixtime and server.mstime are cached less precise timestamps
      that we use every time we don't need an accurate time representation and
      a syscall would be too slow for the number of calls we require.
      
      Such an example is the initialization and update process of the last
      interaction time with the client, that is used for timeouts.
      
      However rdbLoad() can take some time to load the DB, but at the same
      time it did not updated the time during DB loading. This resulted in the
      bug described in issue #1535, where in the replication process the slave
      loads the DB, creates the redisClient representation of its master, but
      the timestamp is so old that the master, under certain conditions, is
      sensed as already "timed out".
      
      Thanks to @yoav-steinberg and Redis Labs Inc for the bug report and
      analysis.
      85492dcf
  10. 12 2月, 2014 2 次提交
    • A
      AOF write error: retry with a frequency of 1 hz. · 96973a7c
      antirez 提交于
      96973a7c
    • A
      AOF: don't abort on write errors unless fsync is 'always'. · fadbbdd3
      antirez 提交于
      A system similar to the RDB write error handling is used, in which when
      we can't write to the AOF file, writes are no longer accepted until we
      are able to write again.
      
      For fsync == always we still abort on errors since there is currently no
      easy way to avoid replying with success to the user otherwise, and this
      would violate the contract with the user of only acknowledging data
      already secured on disk.
      fadbbdd3
  11. 08 2月, 2014 1 次提交
  12. 07 2月, 2014 1 次提交
  13. 03 2月, 2014 1 次提交
  14. 31 1月, 2014 2 次提交
  15. 28 1月, 2014 1 次提交
  16. 23 12月, 2013 1 次提交
  17. 19 12月, 2013 1 次提交
  18. 11 12月, 2013 1 次提交
    • A
      Replication: publish the slave_repl_offset when disconnected from master. · 9a8ae5a5
      antirez 提交于
      When a slave was disconnected from its master the replication offset was
      reported as -1. Now it is reported as the replication offset of the
      previous master, so that failover can be performed using this value in
      order to try to select a slave with more processed data from a set of
      slaves of the old master.
      9a8ae5a5