flexcop-pci.c 12.0 KB
Newer Older
1
/*
2 3 4
 * Linux driver the digital TV devices equipped with B2C2 FlexcopII(b)/III
 * flexcop-pci.c - covers the PCI part including DMA transfers
 * see flexcop.c for copyright information
5 6 7 8 9
 */

#define FC_LOG_PREFIX "flexcop-pci"
#include "flexcop-common.h"

10
static int enable_pid_filtering = 1;
11
module_param(enable_pid_filtering, int, 0444);
12 13
MODULE_PARM_DESC(enable_pid_filtering,
	"enable hardware pid filtering: supported values: 0 (fullts), 1");
14

15
static int irq_chk_intv = 100;
16
module_param(irq_chk_intv, int, 0644);
17
MODULE_PARM_DESC(irq_chk_intv, "set the interval for IRQ streaming watchdog.");
18

19 20 21 22 23 24 25 26 27
#ifdef CONFIG_DVB_B2C2_FLEXCOP_DEBUG
#define dprintk(level,args...) \
	do { if ((debug & level)) printk(args); } while (0)
#define DEBSTATUS ""
#else
#define dprintk(level,args...)
#define DEBSTATUS " (debugging is not enabled)"
#endif

28 29 30 31 32
#define deb_info(args...) dprintk(0x01, args)
#define deb_reg(args...) dprintk(0x02, args)
#define deb_ts(args...) dprintk(0x04, args)
#define deb_irq(args...) dprintk(0x08, args)
#define deb_chk(args...) dprintk(0x10, args)
33

34
static int debug;
35
module_param(debug, int, 0644);
36 37
MODULE_PARM_DESC(debug,
	"set debug level (1=info,2=regs,4=TS,8=irqdma,16=check (|-able))."
38
	DEBSTATUS);
39 40

#define DRIVER_VERSION "0.1"
41
#define DRIVER_NAME "flexcop-pci"
42 43 44 45 46 47 48 49 50 51 52
#define DRIVER_AUTHOR "Patrick Boettcher <patrick.boettcher@desy.de>"

struct flexcop_pci {
	struct pci_dev *pdev;

#define FC_PCI_INIT     0x01
#define FC_PCI_DMA_INIT 0x02
	int init_state;

	void __iomem *io_mem;
	u32 irq;
53 54
	/* buffersize (at least for DMA1, need to be % 188 == 0,
	 * this logic is required */
55 56 57 58 59
#define FC_DEFAULT_DMA1_BUFSIZE (1280 * 188)
#define FC_DEFAULT_DMA2_BUFSIZE (10 * 188)
	struct flexcop_dma dma[2];

	int active_dma1_addr; /* 0 = addr0 of dma1; 1 = addr1 of dma1 */
60
	u32 last_dma1_cur_pos;
L
Lucas De Marchi 已提交
61
	/* position of the pointer last time the timer/packet irq occurred */
62
	int count;
63 64
	int count_prev;
	int stream_problem;
65

66
	spinlock_t irq_lock;
67 68
	unsigned long last_irq;

D
David Howells 已提交
69
	struct delayed_work irq_check_work;
70 71 72
	struct flexcop_device *fc_dev;
};

73
static int lastwreg, lastwval, lastrreg, lastrval;
74

75 76
static flexcop_ibi_value flexcop_pci_read_ibi_reg(struct flexcop_device *fc,
		flexcop_ibi_register r)
77 78 79 80 81 82 83
{
	struct flexcop_pci *fc_pci = fc->bus_specific;
	flexcop_ibi_value v;
	v.raw = readl(fc_pci->io_mem + r);

	if (lastrreg != r || lastrval != v.raw) {
		lastrreg = r; lastrval = v.raw;
84
		deb_reg("new rd: %3x: %08x\n", r, v.raw);
85 86 87 88 89
	}

	return v;
}

90 91
static int flexcop_pci_write_ibi_reg(struct flexcop_device *fc,
		flexcop_ibi_register r, flexcop_ibi_value v)
92 93 94 95 96
{
	struct flexcop_pci *fc_pci = fc->bus_specific;

	if (lastwreg != r || lastwval != v.raw) {
		lastwreg = r; lastwval = v.raw;
97
		deb_reg("new wr: %3x: %08x\n", r, v.raw);
98 99 100 101 102 103
	}

	writel(v.raw, fc_pci->io_mem + r);
	return 0;
}

D
David Howells 已提交
104
static void flexcop_pci_irq_check_work(struct work_struct *work)
105
{
D
David Howells 已提交
106 107
	struct flexcop_pci *fc_pci =
		container_of(work, struct flexcop_pci, irq_check_work.work);
108 109
	struct flexcop_device *fc = fc_pci->fc_dev;

110 111 112 113 114 115
	if (fc->feedcount) {

		if (fc_pci->count == fc_pci->count_prev) {
			deb_chk("no IRQ since the last check\n");
			if (fc_pci->stream_problem++ == 3) {
				struct dvb_demux_feed *feed;
116
				deb_info("flexcop-pci: stream problem, resetting pid filter\n");
117 118 119

				spin_lock_irq(&fc->demux.lock);
				list_for_each_entry(feed, &fc->demux.feed_list,
120
						list_head) {
121 122 123 124
					flexcop_pid_feed_control(fc, feed, 0);
				}

				list_for_each_entry(feed, &fc->demux.feed_list,
125
						list_head) {
126 127 128 129 130 131 132 133 134 135 136
					flexcop_pid_feed_control(fc, feed, 1);
				}
				spin_unlock_irq(&fc->demux.lock);

				fc_pci->stream_problem = 0;
			}
		} else {
			fc_pci->stream_problem = 0;
			fc_pci->count_prev = fc_pci->count;
		}
	}
137 138 139 140 141

	schedule_delayed_work(&fc_pci->irq_check_work,
			msecs_to_jiffies(irq_chk_intv < 100 ? 100 : irq_chk_intv));
}

142 143 144
/* When PID filtering is turned on, we use the timer IRQ, because small amounts
 * of data need to be passed to the user space instantly as well. When PID
 * filtering is turned off, we use the page-change-IRQ */
145
static irqreturn_t flexcop_pci_isr(int irq, void *dev_id)
146 147 148
{
	struct flexcop_pci *fc_pci = dev_id;
	struct flexcop_device *fc = fc_pci->fc_dev;
149
	unsigned long flags;
150
	flexcop_ibi_value v;
151 152
	irqreturn_t ret = IRQ_HANDLED;

153 154
	spin_lock_irqsave(&fc_pci->irq_lock, flags);
	v = fc->read_ibi_reg(fc, irq_20c);
155

156
	/* errors */
157 158 159 160 161 162 163 164 165 166
	if (v.irq_20c.Data_receiver_error)
		deb_chk("data receiver error\n");
	if (v.irq_20c.Continuity_error_flag)
		deb_chk("Contunuity error flag is set\n");
	if (v.irq_20c.LLC_SNAP_FLAG_set)
		deb_chk("LLC_SNAP_FLAG_set is set\n");
	if (v.irq_20c.Transport_Error)
		deb_chk("Transport error\n");

	if ((fc_pci->count % 1000) == 0)
167
		deb_chk("%d valid irq took place so far\n", fc_pci->count);
168

169 170
	if (v.irq_20c.DMA1_IRQ_Status == 1) {
		if (fc_pci->active_dma1_addr == 0)
171 172 173
			flexcop_pass_dmx_packets(fc_pci->fc_dev,
					fc_pci->dma[0].cpu_addr0,
					fc_pci->dma[0].size / 188);
174
		else
175 176 177
			flexcop_pass_dmx_packets(fc_pci->fc_dev,
					fc_pci->dma[0].cpu_addr1,
					fc_pci->dma[0].size / 188);
178 179 180 181 182

		deb_irq("page change to page: %d\n",!fc_pci->active_dma1_addr);
		fc_pci->active_dma1_addr = !fc_pci->active_dma1_addr;
		/* for the timer IRQ we only can use buffer dmx feeding, because we don't have
		 * complete TS packets when reading from the DMA memory */
183
	} else if (v.irq_20c.DMA1_Timer_Status == 1) {
184 185 186 187
		dma_addr_t cur_addr =
			fc->read_ibi_reg(fc,dma1_008).dma_0x8.dma_cur_addr << 2;
		u32 cur_pos = cur_addr - fc_pci->dma[0].dma_addr0;

188 189
		deb_irq("%u irq: %08x cur_addr: %llx: cur_pos: %08x, "
			"last_cur_pos: %08x ",
190 191 192
				jiffies_to_usecs(jiffies - fc_pci->last_irq),
				v.raw, (unsigned long long)cur_addr, cur_pos,
				fc_pci->last_dma1_cur_pos);
193
		fc_pci->last_irq = jiffies;
194

195 196 197 198
		/* buffer end was reached, restarted from the beginning
		 * pass the data from last_cur_pos to the buffer end to the demux
		 */
		if (cur_pos < fc_pci->last_dma1_cur_pos) {
199 200 201
			deb_irq(" end was reached: passing %d bytes ",
				(fc_pci->dma[0].size*2 - 1) -
				fc_pci->last_dma1_cur_pos);
202
			flexcop_pass_dmx_data(fc_pci->fc_dev,
203 204 205 206
				fc_pci->dma[0].cpu_addr0 +
					fc_pci->last_dma1_cur_pos,
				(fc_pci->dma[0].size*2) -
					fc_pci->last_dma1_cur_pos);
207 208 209 210
			fc_pci->last_dma1_cur_pos = 0;
		}

		if (cur_pos > fc_pci->last_dma1_cur_pos) {
211 212
			deb_irq(" passing %d bytes ",
				cur_pos - fc_pci->last_dma1_cur_pos);
213
			flexcop_pass_dmx_data(fc_pci->fc_dev,
214 215 216
				fc_pci->dma[0].cpu_addr0 +
					fc_pci->last_dma1_cur_pos,
				cur_pos - fc_pci->last_dma1_cur_pos);
217
		}
218
		deb_irq("\n");
219 220

		fc_pci->last_dma1_cur_pos = cur_pos;
221 222
		fc_pci->count++;
	} else {
223 224
		deb_irq("isr for flexcop called, "
			"apparently without reason (%08x)\n", v.raw);
225
		ret = IRQ_NONE;
226
	}
227

228
	spin_unlock_irqrestore(&fc_pci->irq_lock, flags);
229
	return ret;
230 231 232 233 234 235
}

static int flexcop_pci_stream_control(struct flexcop_device *fc, int onoff)
{
	struct flexcop_pci *fc_pci = fc->bus_specific;
	if (onoff) {
236 237 238 239 240
		flexcop_dma_config(fc, &fc_pci->dma[0], FC_DMA_1);
		flexcop_dma_config(fc, &fc_pci->dma[1], FC_DMA_2);
		flexcop_dma_config_timer(fc, FC_DMA_1, 0);
		flexcop_dma_xfer_control(fc, FC_DMA_1,
				FC_DMA_SUBADDR_0 | FC_DMA_SUBADDR_1, 1);
241
		deb_irq("DMA xfer enabled\n");
242

243
		fc_pci->last_dma1_cur_pos = 0;
244
		flexcop_dma_control_timer_irq(fc, FC_DMA_1, 1);
245
		deb_irq("IRQ enabled\n");
246
		fc_pci->count_prev = fc_pci->count;
247
	} else {
248
		flexcop_dma_control_timer_irq(fc, FC_DMA_1, 0);
249
		deb_irq("IRQ disabled\n");
250

251 252
		flexcop_dma_xfer_control(fc, FC_DMA_1,
			 FC_DMA_SUBADDR_0 | FC_DMA_SUBADDR_1, 0);
253
		deb_irq("DMA xfer disabled\n");
254 255 256 257 258 259 260
	}
	return 0;
}

static int flexcop_pci_dma_init(struct flexcop_pci *fc_pci)
{
	int ret;
261 262 263
	ret = flexcop_dma_allocate(fc_pci->pdev, &fc_pci->dma[0],
			FC_DEFAULT_DMA1_BUFSIZE);
	if (ret != 0)
264 265
		return ret;

266 267 268
	ret = flexcop_dma_allocate(fc_pci->pdev, &fc_pci->dma[1],
			FC_DEFAULT_DMA2_BUFSIZE);
	if (ret != 0) {
269 270 271
		flexcop_dma_free(&fc_pci->dma[0]);
		return ret;
	}
272

273 274 275 276
	flexcop_sram_set_dest(fc_pci->fc_dev, FC_SRAM_DEST_MEDIA |
			FC_SRAM_DEST_NET, FC_SRAM_DEST_TARGET_DMA1);
	flexcop_sram_set_dest(fc_pci->fc_dev, FC_SRAM_DEST_CAO |
			FC_SRAM_DEST_CAI, FC_SRAM_DEST_TARGET_DMA2);
277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293
	fc_pci->init_state |= FC_PCI_DMA_INIT;
	return ret;
}

static void flexcop_pci_dma_exit(struct flexcop_pci *fc_pci)
{
	if (fc_pci->init_state & FC_PCI_DMA_INIT) {
		flexcop_dma_free(&fc_pci->dma[0]);
		flexcop_dma_free(&fc_pci->dma[1]);
	}
	fc_pci->init_state &= ~FC_PCI_DMA_INIT;
}

static int flexcop_pci_init(struct flexcop_pci *fc_pci)
{
	int ret;

B
Bjørn Mork 已提交
294
	info("card revision %x", fc_pci->pdev->revision);
295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311

	if ((ret = pci_enable_device(fc_pci->pdev)) != 0)
		return ret;
	pci_set_master(fc_pci->pdev);

	if ((ret = pci_request_regions(fc_pci->pdev, DRIVER_NAME)) != 0)
		goto err_pci_disable_device;

	fc_pci->io_mem = pci_iomap(fc_pci->pdev, 0, 0x800);

	if (!fc_pci->io_mem) {
		err("cannot map io memory\n");
		ret = -EIO;
		goto err_pci_release_regions;
	}

	pci_set_drvdata(fc_pci->pdev, fc_pci);
312
	spin_lock_init(&fc_pci->irq_lock);
313
	if ((ret = request_irq(fc_pci->pdev->irq, flexcop_pci_isr,
314
					IRQF_SHARED, DRIVER_NAME, fc_pci)) != 0)
315 316 317
		goto err_pci_iounmap;

	fc_pci->init_state |= FC_PCI_INIT;
318
	return ret;
319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341

err_pci_iounmap:
	pci_iounmap(fc_pci->pdev, fc_pci->io_mem);
	pci_set_drvdata(fc_pci->pdev, NULL);
err_pci_release_regions:
	pci_release_regions(fc_pci->pdev);
err_pci_disable_device:
	pci_disable_device(fc_pci->pdev);
	return ret;
}

static void flexcop_pci_exit(struct flexcop_pci *fc_pci)
{
	if (fc_pci->init_state & FC_PCI_INIT) {
		free_irq(fc_pci->pdev->irq, fc_pci);
		pci_iounmap(fc_pci->pdev, fc_pci->io_mem);
		pci_set_drvdata(fc_pci->pdev, NULL);
		pci_release_regions(fc_pci->pdev);
		pci_disable_device(fc_pci->pdev);
	}
	fc_pci->init_state &= ~FC_PCI_INIT;
}

342 343
static int flexcop_pci_probe(struct pci_dev *pdev,
		const struct pci_device_id *ent)
344 345 346 347 348 349 350 351 352 353
{
	struct flexcop_device *fc;
	struct flexcop_pci *fc_pci;
	int ret = -ENOMEM;

	if ((fc = flexcop_device_kmalloc(sizeof(struct flexcop_pci))) == NULL) {
		err("out of memory\n");
		return -ENOMEM;
	}

354
	/* general flexcop init */
355 356 357 358 359 360 361 362 363
	fc_pci = fc->bus_specific;
	fc_pci->fc_dev = fc;

	fc->read_ibi_reg = flexcop_pci_read_ibi_reg;
	fc->write_ibi_reg = flexcop_pci_write_ibi_reg;
	fc->i2c_request = flexcop_i2c_request;
	fc->get_mac_addr = flexcop_eeprom_check_mac_addr;
	fc->stream_control = flexcop_pci_stream_control;

364 365 366 367 368
	if (enable_pid_filtering)
		info("will use the HW PID filter.");
	else
		info("will pass the complete TS to the demuxer.");

369 370 371
	fc->pid_filtering = enable_pid_filtering;
	fc->bus_type = FC_PCI;
	fc->dev = &pdev->dev;
372
	fc->owner = THIS_MODULE;
373

374
	/* bus specific part */
375 376 377 378
	fc_pci->pdev = pdev;
	if ((ret = flexcop_pci_init(fc_pci)) != 0)
		goto err_kfree;

379
	/* init flexcop */
380 381 382
	if ((ret = flexcop_device_initialize(fc)) != 0)
		goto err_pci_exit;

383
	/* init dma */
384 385 386
	if ((ret = flexcop_pci_dma_init(fc_pci)) != 0)
		goto err_fc_exit;

D
David Howells 已提交
387
	INIT_DELAYED_WORK(&fc_pci->irq_check_work, flexcop_pci_irq_check_work);
388

389 390 391 392 393
	if (irq_chk_intv > 0)
		schedule_delayed_work(&fc_pci->irq_check_work,
				msecs_to_jiffies(irq_chk_intv < 100 ?
					100 :
					irq_chk_intv));
394 395
	return ret;

396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411
err_fc_exit:
	flexcop_device_exit(fc);
err_pci_exit:
	flexcop_pci_exit(fc_pci);
err_kfree:
	flexcop_device_kfree(fc);
	return ret;
}

/* in theory every _exit function should be called exactly two times,
 * here and in the bail-out-part of the _init-function
 */
static void flexcop_pci_remove(struct pci_dev *pdev)
{
	struct flexcop_pci *fc_pci = pci_get_drvdata(pdev);

412 413 414
	if (irq_chk_intv > 0)
		cancel_delayed_work(&fc_pci->irq_check_work);

415 416 417 418 419 420 421 422 423 424 425 426 427 428
	flexcop_pci_dma_exit(fc_pci);
	flexcop_device_exit(fc_pci->fc_dev);
	flexcop_pci_exit(fc_pci);
	flexcop_device_kfree(fc_pci->fc_dev);
}

static struct pci_device_id flexcop_pci_tbl[] = {
	{ PCI_DEVICE(0x13d0, 0x2103) },
	{ },
};

MODULE_DEVICE_TABLE(pci, flexcop_pci_tbl);

static struct pci_driver flexcop_pci_driver = {
429
	.name     = "b2c2_flexcop_pci",
430
	.id_table = flexcop_pci_tbl,
431 432
	.probe    = flexcop_pci_probe,
	.remove   = flexcop_pci_remove,
433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450
};

static int __init flexcop_pci_module_init(void)
{
	return pci_register_driver(&flexcop_pci_driver);
}

static void __exit flexcop_pci_module_exit(void)
{
	pci_unregister_driver(&flexcop_pci_driver);
}

module_init(flexcop_pci_module_init);
module_exit(flexcop_pci_module_exit);

MODULE_AUTHOR(DRIVER_AUTHOR);
MODULE_DESCRIPTION(DRIVER_NAME);
MODULE_LICENSE("GPL");