dma.h 10.4 KB
Newer Older
1
/*
2
 * Copyright(c) 2004 - 2009 Intel Corporation. All rights reserved.
3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
 *
 * This program is free software; you can redistribute it and/or modify it
 * under the terms of the GNU General Public License as published by the Free
 * Software Foundation; either version 2 of the License, or (at your option)
 * any later version.
 *
 * This program is distributed in the hope that it will be useful, but WITHOUT
 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
 * more details.
 *
 * You should have received a copy of the GNU General Public License along with
 * this program; if not, write to the Free Software Foundation, Inc., 59
 * Temple Place - Suite 330, Boston, MA  02111-1307, USA.
 *
 * The full GNU General Public License is included in this distribution in the
 * file called COPYING.
 */
#ifndef IOATDMA_H
#define IOATDMA_H

#include <linux/dmaengine.h>
D
Dan Williams 已提交
25
#include "hw.h"
26
#include "registers.h"
27 28 29
#include <linux/init.h>
#include <linux/dmapool.h>
#include <linux/cache.h>
30
#include <linux/pci_ids.h>
31
#include <net/tcp.h>
32

D
Dan Williams 已提交
33
#define IOAT_DMA_VERSION  "4.00"
34

35
#define IOAT_LOW_COMPLETION_MASK	0xffffffc0
36 37
#define IOAT_DMA_DCA_ANY_CPU		~0

D
Dan Williams 已提交
38 39
#define to_ioatdma_device(dev) container_of(dev, struct ioatdma_device, common)
#define to_ioat_desc(lh) container_of(lh, struct ioat_desc_sw, node)
40 41
#define tx_to_ioat_desc(tx) container_of(tx, struct ioat_desc_sw, txd)
#define to_dev(ioat_chan) (&(ioat_chan)->device->pdev->dev)
D
Dan Williams 已提交
42 43 44 45 46 47 48 49 50

#define chan_num(ch) ((int)((ch)->reg_base - (ch)->device->reg_base) / 0x80)

/*
 * workaround for IOAT ver.3.0 null descriptor issue
 * (channel returns error when size is 0)
 */
#define NULL_DESC_BUFFER_SIZE 1

51
/**
52
 * struct ioatdma_device - internal representation of a IOAT device
53 54 55 56
 * @pdev: PCI-Express device
 * @reg_base: MMIO register space base address
 * @dma_pool: for allocating DMA descriptors
 * @common: embedded struct dma_device
57
 * @version: version of ioatdma device
58 59
 * @msix_entries: irq handlers
 * @idx: per channel data
60 61
 * @dca: direct cache access context
 * @intr_quirk: interrupt setup quirk (for ioat_v1 devices)
62
 * @enumerate_channels: hw version specific channel enumeration
63
 * @reset_hw: hw version specific channel (re)initialization
64
 * @cleanup_fn: select between the v2 and v3 cleanup routines
65
 * @timer_fn: select between the v2 and v3 timer watchdog routines
D
Dan Williams 已提交
66
 * @self_test: hardware version specific self test for each supported op type
67 68
 *
 * Note: the v3 cleanup routine supports raid operations
69
 */
70
struct ioatdma_device {
71
	struct pci_dev *pdev;
A
Al Viro 已提交
72
	void __iomem *reg_base;
73 74 75
	struct pci_pool *dma_pool;
	struct pci_pool *completion_pool;
	struct dma_device common;
76
	u8 version;
77
	struct msix_entry msix_entries[4];
78
	struct ioat_chan_common *idx[4];
79 80
	struct dca_provider *dca;
	void (*intr_quirk)(struct ioatdma_device *device);
81
	int (*enumerate_channels)(struct ioatdma_device *device);
82
	int (*reset_hw)(struct ioat_chan_common *chan);
83
	void (*cleanup_fn)(unsigned long data);
84
	void (*timer_fn)(unsigned long data);
D
Dan Williams 已提交
85
	int (*self_test)(struct ioatdma_device *device);
86 87
};

88
struct ioat_chan_common {
89
	struct dma_chan common;
A
Al Viro 已提交
90
	void __iomem *reg_base;
91 92
	unsigned long last_completion;
	spinlock_t cleanup_lock;
93
	dma_cookie_t completed_cookie;
94 95 96 97
	unsigned long state;
	#define IOAT_COMPLETION_PENDING 0
	#define IOAT_COMPLETION_ACK 1
	#define IOAT_RESET_PENDING 2
D
Dan Williams 已提交
98
	#define IOAT_KOBJ_INIT_FAIL 3
99 100
	struct timer_list timer;
	#define COMPLETION_TIMEOUT msecs_to_jiffies(100)
101
	#define IDLE_TIMEOUT msecs_to_jiffies(2000)
102
	#define RESET_DELAY msecs_to_jiffies(100)
103
	struct ioatdma_device *device;
104 105
	dma_addr_t completion_dma;
	u64 *completion;
106
	struct tasklet_struct cleanup_task;
D
Dan Williams 已提交
107
	struct kobject kobj;
108 109
};

D
Dan Williams 已提交
110 111 112 113
struct ioat_sysfs_entry {
	struct attribute attr;
	ssize_t (*show)(struct dma_chan *, char *);
};
114

115 116 117 118 119 120 121 122 123 124 125 126 127 128
/**
 * struct ioat_dma_chan - internal representation of a DMA channel
 */
struct ioat_dma_chan {
	struct ioat_chan_common base;

	size_t xfercap;	/* XFERCAP register value expanded out */

	spinlock_t desc_lock;
	struct list_head free_desc;
	struct list_head used_desc;

	int pending;
	u16 desccount;
D
Dan Williams 已提交
129
	u16 active;
130 131 132 133 134 135 136 137 138 139 140 141 142 143
};

static inline struct ioat_chan_common *to_chan_common(struct dma_chan *c)
{
	return container_of(c, struct ioat_chan_common, common);
}

static inline struct ioat_dma_chan *to_ioat_chan(struct dma_chan *c)
{
	struct ioat_chan_common *chan = to_chan_common(c);

	return container_of(chan, struct ioat_dma_chan, base);
}

144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169
/**
 * ioat_is_complete - poll the status of an ioat transaction
 * @c: channel handle
 * @cookie: transaction identifier
 * @done: if set, updated with last completed transaction
 * @used: if set, updated with last used transaction
 */
static inline enum dma_status
ioat_is_complete(struct dma_chan *c, dma_cookie_t cookie,
		 dma_cookie_t *done, dma_cookie_t *used)
{
	struct ioat_chan_common *chan = to_chan_common(c);
	dma_cookie_t last_used;
	dma_cookie_t last_complete;

	last_used = c->cookie;
	last_complete = chan->completed_cookie;

	if (done)
		*done = last_complete;
	if (used)
		*used = last_used;

	return dma_async_is_complete(cookie, last_complete, last_used);
}

170 171 172 173
/* wrapper around hardware descriptor format + additional software fields */

/**
 * struct ioat_desc_sw - wrapper around hardware descriptor
174
 * @hw: hardware DMA descriptor (for memcpy)
175
 * @node: this descriptor will either be on the free list,
D
Dan Williams 已提交
176
 *     or attached to a transaction list (tx_list)
177
 * @txd: the generic software descriptor for all engines
D
Dan Williams 已提交
178
 * @id: identifier for debug
179 180 181 182
 */
struct ioat_desc_sw {
	struct ioat_dma_descriptor *hw;
	struct list_head node;
183
	size_t len;
D
Dan Williams 已提交
184
	struct list_head tx_list;
185
	struct dma_async_tx_descriptor txd;
D
Dan Williams 已提交
186 187 188
	#ifdef DEBUG
	int id;
	#endif
189 190
};

D
Dan Williams 已提交
191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214
#ifdef DEBUG
#define set_desc_id(desc, i) ((desc)->id = (i))
#define desc_id(desc) ((desc)->id)
#else
#define set_desc_id(desc, i)
#define desc_id(desc) (0)
#endif

static inline void
__dump_desc_dbg(struct ioat_chan_common *chan, struct ioat_dma_descriptor *hw,
		struct dma_async_tx_descriptor *tx, int id)
{
	struct device *dev = to_dev(chan);

	dev_dbg(dev, "desc[%d]: (%#llx->%#llx) cookie: %d flags: %#x"
		" ctl: %#x (op: %d int_en: %d compl: %d)\n", id,
		(unsigned long long) tx->phys,
		(unsigned long long) hw->next, tx->cookie, tx->flags,
		hw->ctl, hw->ctl_f.op, hw->ctl_f.int_en, hw->ctl_f.compl_write);
}

#define dump_desc_dbg(c, d) \
	({ if (d) __dump_desc_dbg(&c->base, d->hw, &d->txd, desc_id(d)); 0; })

215
static inline void ioat_set_tcp_copy_break(unsigned long copybreak)
216 217
{
	#ifdef CONFIG_NET_DMA
218
	sysctl_tcp_dma_copybreak = copybreak;
219 220 221
	#endif
}

222 223 224 225 226 227
static inline struct ioat_chan_common *
ioat_chan_by_index(struct ioatdma_device *device, int index)
{
	return device->idx[index];
}

228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268
static inline u64 ioat_chansts(struct ioat_chan_common *chan)
{
	u8 ver = chan->device->version;
	u64 status;
	u32 status_lo;

	/* We need to read the low address first as this causes the
	 * chipset to latch the upper bits for the subsequent read
	 */
	status_lo = readl(chan->reg_base + IOAT_CHANSTS_OFFSET_LOW(ver));
	status = readl(chan->reg_base + IOAT_CHANSTS_OFFSET_HIGH(ver));
	status <<= 32;
	status |= status_lo;

	return status;
}

static inline void ioat_start(struct ioat_chan_common *chan)
{
	u8 ver = chan->device->version;

	writeb(IOAT_CHANCMD_START, chan->reg_base + IOAT_CHANCMD_OFFSET(ver));
}

static inline u64 ioat_chansts_to_addr(u64 status)
{
	return status & IOAT_CHANSTS_COMPLETED_DESCRIPTOR_ADDR;
}

static inline u32 ioat_chanerr(struct ioat_chan_common *chan)
{
	return readl(chan->reg_base + IOAT_CHANERR_OFFSET);
}

static inline void ioat_suspend(struct ioat_chan_common *chan)
{
	u8 ver = chan->device->version;

	writeb(IOAT_CHANCMD_SUSPEND, chan->reg_base + IOAT_CHANCMD_OFFSET(ver));
}

269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284
static inline void ioat_reset(struct ioat_chan_common *chan)
{
	u8 ver = chan->device->version;

	writeb(IOAT_CHANCMD_RESET, chan->reg_base + IOAT_CHANCMD_OFFSET(ver));
}

static inline bool ioat_reset_pending(struct ioat_chan_common *chan)
{
	u8 ver = chan->device->version;
	u8 cmd;

	cmd = readb(chan->reg_base + IOAT_CHANCMD_OFFSET(ver));
	return (cmd & IOAT_CHANCMD_RESET) == IOAT_CHANCMD_RESET;
}

285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317
static inline void ioat_set_chainaddr(struct ioat_dma_chan *ioat, u64 addr)
{
	struct ioat_chan_common *chan = &ioat->base;

	writel(addr & 0x00000000FFFFFFFF,
	       chan->reg_base + IOAT1_CHAINADDR_OFFSET_LOW);
	writel(addr >> 32,
	       chan->reg_base + IOAT1_CHAINADDR_OFFSET_HIGH);
}

static inline bool is_ioat_active(unsigned long status)
{
	return ((status & IOAT_CHANSTS_STATUS) == IOAT_CHANSTS_ACTIVE);
}

static inline bool is_ioat_idle(unsigned long status)
{
	return ((status & IOAT_CHANSTS_STATUS) == IOAT_CHANSTS_DONE);
}

static inline bool is_ioat_halted(unsigned long status)
{
	return ((status & IOAT_CHANSTS_STATUS) == IOAT_CHANSTS_HALTED);
}

static inline bool is_ioat_suspended(unsigned long status)
{
	return ((status & IOAT_CHANSTS_STATUS) == IOAT_CHANSTS_SUSPENDED);
}

/* channel was fatally programmed */
static inline bool is_ioat_bug(unsigned long err)
{
318
	return !!err;
319 320
}

321 322 323 324 325 326 327 328 329 330
static inline void ioat_unmap(struct pci_dev *pdev, dma_addr_t addr, size_t len,
			      int direction, enum dma_ctrl_flags flags, bool dst)
{
	if ((dst && (flags & DMA_COMPL_DEST_UNMAP_SINGLE)) ||
	    (!dst && (flags & DMA_COMPL_SRC_UNMAP_SINGLE)))
		pci_unmap_single(pdev, addr, len, direction);
	else
		pci_unmap_page(pdev, addr, len, direction);
}

331 332 333
int __devinit ioat_probe(struct ioatdma_device *device);
int __devinit ioat_register(struct ioatdma_device *device);
int __devinit ioat1_dma_probe(struct ioatdma_device *dev, int dca);
D
Dan Williams 已提交
334
int __devinit ioat_dma_self_test(struct ioatdma_device *device);
335 336 337
void __devexit ioat_dma_remove(struct ioatdma_device *device);
struct dca_provider * __devinit ioat_dca_init(struct pci_dev *pdev,
					      void __iomem *iobase);
338 339
unsigned long ioat_get_current_completion(struct ioat_chan_common *chan);
void ioat_init_channel(struct ioatdma_device *device,
340 341 342
		       struct ioat_chan_common *chan, int idx);
enum dma_status ioat_is_dma_complete(struct dma_chan *c, dma_cookie_t cookie,
				     dma_cookie_t *done, dma_cookie_t *used);
343 344
void ioat_dma_unmap(struct ioat_chan_common *chan, enum dma_ctrl_flags flags,
		    size_t len, struct ioat_dma_descriptor *hw);
345 346
bool ioat_cleanup_preamble(struct ioat_chan_common *chan,
			   unsigned long *phys_complete);
D
Dan Williams 已提交
347 348
void ioat_kobject_add(struct ioatdma_device *device, struct kobj_type *type);
void ioat_kobject_del(struct ioatdma_device *device);
349
extern const struct sysfs_ops ioat_sysfs_ops;
D
Dan Williams 已提交
350 351
extern struct ioat_sysfs_entry ioat_version_attr;
extern struct ioat_sysfs_entry ioat_cap_attr;
352
#endif /* IOATDMA_H */