提交 fe4ada2d 编写于 作者: R Randy Dunlap 提交者: David S. Miller

[IOAT]: fix header file kernel-doc

Fix kernel-doc problems in include/linux/dmaengine.h:
- add some fields/parameters
- expand some descriptions
- fix typos
Signed-off-by: NRandy Dunlap <rdunlap@xenotime.net>
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
上级 b0026624
...@@ -44,7 +44,7 @@ enum dma_event { ...@@ -44,7 +44,7 @@ enum dma_event {
}; };
/** /**
* typedef dma_cookie_t * typedef dma_cookie_t - an opaque DMA cookie
* *
* if dma_cookie_t is >0 it's a DMA request cookie, <0 it's an error code * if dma_cookie_t is >0 it's a DMA request cookie, <0 it's an error code
*/ */
...@@ -80,14 +80,14 @@ struct dma_chan_percpu { ...@@ -80,14 +80,14 @@ struct dma_chan_percpu {
/** /**
* struct dma_chan - devices supply DMA channels, clients use them * struct dma_chan - devices supply DMA channels, clients use them
* @client: ptr to the client user of this chan, will be NULL when unused * @client: ptr to the client user of this chan, will be %NULL when unused
* @device: ptr to the dma device who supplies this channel, always !NULL * @device: ptr to the dma device who supplies this channel, always !%NULL
* @cookie: last cookie value returned to client * @cookie: last cookie value returned to client
* @chan_id: * @chan_id: channel ID for sysfs
* @class_dev: * @class_dev: class device for sysfs
* @refcount: kref, used in "bigref" slow-mode * @refcount: kref, used in "bigref" slow-mode
* @slow_ref: * @slow_ref: indicates that the DMA channel is free
* @rcu: * @rcu: the DMA channel's RCU head
* @client_node: used to add this to the client chan list * @client_node: used to add this to the client chan list
* @device_node: used to add this to the device chan list * @device_node: used to add this to the device chan list
* @local: per-cpu pointer to a struct dma_chan_percpu * @local: per-cpu pointer to a struct dma_chan_percpu
...@@ -162,10 +162,17 @@ struct dma_client { ...@@ -162,10 +162,17 @@ struct dma_client {
* @chancnt: how many DMA channels are supported * @chancnt: how many DMA channels are supported
* @channels: the list of struct dma_chan * @channels: the list of struct dma_chan
* @global_node: list_head for global dma_device_list * @global_node: list_head for global dma_device_list
* @refcount: * @refcount: reference count
* @done: * @done: IO completion struct
* @dev_id: * @dev_id: unique device ID
* Other func ptrs: used to make use of this device's capabilities * @device_alloc_chan_resources: allocate resources and return the
* number of allocated descriptors
* @device_free_chan_resources: release DMA channel's resources
* @device_memcpy_buf_to_buf: memcpy buf pointer to buf pointer
* @device_memcpy_buf_to_pg: memcpy buf pointer to struct page
* @device_memcpy_pg_to_pg: memcpy struct page/offset to struct page/offset
* @device_memcpy_complete: poll the status of an IOAT DMA transaction
* @device_memcpy_issue_pending: push appended descriptors to hardware
*/ */
struct dma_device { struct dma_device {
...@@ -211,7 +218,7 @@ void dma_async_client_chan_request(struct dma_client *client, ...@@ -211,7 +218,7 @@ void dma_async_client_chan_request(struct dma_client *client,
* Both @dest and @src must be mappable to a bus address according to the * Both @dest and @src must be mappable to a bus address according to the
* DMA mapping API rules for streaming mappings. * DMA mapping API rules for streaming mappings.
* Both @dest and @src must stay memory resident (kernel memory or locked * Both @dest and @src must stay memory resident (kernel memory or locked
* user space pages) * user space pages).
*/ */
static inline dma_cookie_t dma_async_memcpy_buf_to_buf(struct dma_chan *chan, static inline dma_cookie_t dma_async_memcpy_buf_to_buf(struct dma_chan *chan,
void *dest, void *src, size_t len) void *dest, void *src, size_t len)
...@@ -225,7 +232,7 @@ static inline dma_cookie_t dma_async_memcpy_buf_to_buf(struct dma_chan *chan, ...@@ -225,7 +232,7 @@ static inline dma_cookie_t dma_async_memcpy_buf_to_buf(struct dma_chan *chan,
} }
/** /**
* dma_async_memcpy_buf_to_pg - offloaded copy * dma_async_memcpy_buf_to_pg - offloaded copy from address to page
* @chan: DMA channel to offload copy to * @chan: DMA channel to offload copy to
* @page: destination page * @page: destination page
* @offset: offset in page to copy to * @offset: offset in page to copy to
...@@ -250,18 +257,18 @@ static inline dma_cookie_t dma_async_memcpy_buf_to_pg(struct dma_chan *chan, ...@@ -250,18 +257,18 @@ static inline dma_cookie_t dma_async_memcpy_buf_to_pg(struct dma_chan *chan,
} }
/** /**
* dma_async_memcpy_buf_to_pg - offloaded copy * dma_async_memcpy_pg_to_pg - offloaded copy from page to page
* @chan: DMA channel to offload copy to * @chan: DMA channel to offload copy to
* @dest_page: destination page * @dest_pg: destination page
* @dest_off: offset in page to copy to * @dest_off: offset in page to copy to
* @src_page: source page * @src_pg: source page
* @src_off: offset in page to copy from * @src_off: offset in page to copy from
* @len: length * @len: length
* *
* Both @dest_page/@dest_off and @src_page/@src_off must be mappable to a bus * Both @dest_page/@dest_off and @src_page/@src_off must be mappable to a bus
* address according to the DMA mapping API rules for streaming mappings. * address according to the DMA mapping API rules for streaming mappings.
* Both @dest_page/@dest_off and @src_page/@src_off must stay memory resident * Both @dest_page/@dest_off and @src_page/@src_off must stay memory resident
* (kernel memory or locked user space pages) * (kernel memory or locked user space pages).
*/ */
static inline dma_cookie_t dma_async_memcpy_pg_to_pg(struct dma_chan *chan, static inline dma_cookie_t dma_async_memcpy_pg_to_pg(struct dma_chan *chan,
struct page *dest_pg, unsigned int dest_off, struct page *src_pg, struct page *dest_pg, unsigned int dest_off, struct page *src_pg,
...@@ -278,7 +285,7 @@ static inline dma_cookie_t dma_async_memcpy_pg_to_pg(struct dma_chan *chan, ...@@ -278,7 +285,7 @@ static inline dma_cookie_t dma_async_memcpy_pg_to_pg(struct dma_chan *chan,
/** /**
* dma_async_memcpy_issue_pending - flush pending copies to HW * dma_async_memcpy_issue_pending - flush pending copies to HW
* @chan: * @chan: target DMA channel
* *
* This allows drivers to push copies to HW in batches, * This allows drivers to push copies to HW in batches,
* reducing MMIO writes where possible. * reducing MMIO writes where possible.
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册