1. 08 5月, 2012 1 次提交
  2. 21 10月, 2010 2 次提交
    • Y
      ceph: factor out libceph from Ceph file system · 3d14c5d2
      Yehuda Sadeh 提交于
      This factors out protocol and low-level storage parts of ceph into a
      separate libceph module living in net/ceph and include/linux/ceph.  This
      is mostly a matter of moving files around.  However, a few key pieces
      of the interface change as well:
      
       - ceph_client becomes ceph_fs_client and ceph_client, where the latter
         captures the mon and osd clients, and the fs_client gets the mds client
         and file system specific pieces.
       - Mount option parsing and debugfs setup is correspondingly broken into
         two pieces.
       - The mon client gets a generic handler callback for otherwise unknown
         messages (mds map, in this case).
       - The basic supported/required feature bits can be expanded (and are by
         ceph_fs_client).
      
      No functional change, aside from some subtle error handling cases that got
      cleaned up in the refactoring process.
      Signed-off-by: NSage Weil <sage@newdream.net>
      3d14c5d2
    • Y
      ceph: lookup pool in osdmap by name · 7669a2c9
      Yehuda Sadeh 提交于
      Implement a pool lookup by name.  This will be used by rbd.
      Signed-off-by: NYehuda Sadeh <yehuda@hq.newdream.net>
      Signed-off-by: NSage Weil <sage@newdream.net>
      7669a2c9
  3. 12 5月, 2010 1 次提交
  4. 10 4月, 2010 1 次提交
  5. 18 2月, 2010 1 次提交
    • S
      ceph: use rbtree for pg pools; decode new osdmap format · 4fc51be8
      Sage Weil 提交于
      Since we can now create and destroy pg pools, the pool ids will be sparse,
      and an array no longer makes sense for looking up by pool id.  Use an
      rbtree instead.
      
      The OSDMap encoding also no longer has a max pool count (previously used to
      allocate the array).  There is a new pool_max, that is the largest pool id
      we've ever used, although we don't actually need it in the client.
      Signed-off-by: NSage Weil <sage@newdream.net>
      4fc51be8
  6. 05 11月, 2009 1 次提交
  7. 07 10月, 2009 1 次提交
    • S
      ceph: OSD client · f24e9980
      Sage Weil 提交于
      The OSD client is responsible for reading and writing data from/to the
      object storage pool.  This includes determining where objects are
      stored in the cluster, and ensuring that requests are retried or
      redirected in the event of a node failure or data migration.
      
      If an OSD does not respond before a timeout expires, keepalive
      messages are sent across the lossless, ordered communications channel
      to ensure that any break in the TCP is discovered.  If the session
      does reset, a reconnection is attempted and affected requests are
      resent (by the message transport layer).
      Signed-off-by: NSage Weil <sage@newdream.net>
      f24e9980