cx231xx-vbi.c 17.6 KB
Newer Older
1 2 3 4
/*
   cx231xx_vbi.c - driver for Conexant Cx23100/101/102 USB video capture devices

   Copyright (C) 2008 <srinivasa.deevi at conexant dot com>
5
	Based on cx88 driver
6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30

   This program is free software; you can redistribute it and/or modify
   it under the terms of the GNU General Public License as published by
   the Free Software Foundation; either version 2 of the License, or
   (at your option) any later version.

   This program is distributed in the hope that it will be useful,
   but WITHOUT ANY WARRANTY; without even the implied warranty of
   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
   GNU General Public License for more details.

   You should have received a copy of the GNU General Public License
   along with this program; if not, write to the Free Software
   Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
 */

#include <linux/init.h>
#include <linux/list.h>
#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/bitmap.h>
#include <linux/usb.h>
#include <linux/i2c.h>
#include <linux/mm.h>
#include <linux/mutex.h>
31
#include <linux/slab.h>
32 33 34 35 36 37 38 39 40 41

#include <media/v4l2-common.h>
#include <media/v4l2-ioctl.h>
#include <media/v4l2-chip-ident.h>
#include <media/msp3400.h>
#include <media/tuner.h>

#include "cx231xx.h"
#include "cx231xx-vbi.h"

42
static inline void print_err_status(struct cx231xx *dev, int packet, int status)
43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72
{
	char *errmsg = "Unknown";

	switch (status) {
	case -ENOENT:
		errmsg = "unlinked synchronuously";
		break;
	case -ECONNRESET:
		errmsg = "unlinked asynchronuously";
		break;
	case -ENOSR:
		errmsg = "Buffer error (overrun)";
		break;
	case -EPIPE:
		errmsg = "Stalled (device not responding)";
		break;
	case -EOVERFLOW:
		errmsg = "Babble (bad cable?)";
		break;
	case -EPROTO:
		errmsg = "Bit-stuff error (bad cable?)";
		break;
	case -EILSEQ:
		errmsg = "CRC/Timeout (could be anything)";
		break;
	case -ETIME:
		errmsg = "Device does not respond";
		break;
	}
	if (packet < 0) {
73 74
		cx231xx_err(DRIVER_NAME "URB status %d [%s].\n", status,
			    errmsg);
75 76
	} else {
		cx231xx_err(DRIVER_NAME "URB packet %d, status %d [%s].\n",
77
			    packet, status, errmsg);
78 79 80 81 82 83 84 85
	}
}

/*
 * Controls the isoc copy of each urb packet
 */
static inline int cx231xx_isoc_vbi_copy(struct cx231xx *dev, struct urb *urb)
{
86 87 88
	struct cx231xx_buffer *buf;
	struct cx231xx_dmaqueue *dma_q = urb->context;
	int rc = 1;
89
	unsigned char *p_buffer;
90 91
	u32 bytes_parsed = 0, buffer_size = 0;
	u8 sav_eav = 0;
92 93 94 95

	if (!dev)
		return 0;

96
	if (dev->state & DEV_DISCONNECTED)
97 98 99 100 101 102 103 104
		return 0;

	if (urb->status < 0) {
		print_err_status(dev, -1, urb->status);
		if (urb->status == -ENOENT)
			return 0;
	}

105
	buf = dev->vbi_mode.bulk_ctl.buf;
106

107 108 109 110 111 112 113 114
	/* get buffer pointer and length */
	p_buffer = urb->transfer_buffer;
	buffer_size = urb->actual_length;

	if (buffer_size > 0) {
		bytes_parsed = 0;

		if (dma_q->is_partial_line) {
115 116
			/* Handle the case where we were working on a partial
			   line */
117 118
			sav_eav = dma_q->last_sav;
		} else {
119 120 121 122
			/* Check for a SAV/EAV overlapping the
			   buffer boundary */

			sav_eav = cx231xx_find_boundary_SAV_EAV(p_buffer,
123 124 125 126 127
							  dma_q->partial_buf,
							  &bytes_parsed);
		}

		sav_eav &= 0xF0;
128 129
		/* Get the first line if we have some portion of an SAV/EAV from
		   the last buffer or a partial line */
130
		if (sav_eav) {
131 132 133 134
			bytes_parsed += cx231xx_get_vbi_line(dev, dma_q,
				sav_eav,		       /* SAV/EAV */
				p_buffer + bytes_parsed,       /* p_buffer */
				buffer_size - bytes_parsed);   /* buffer size */
135 136 137 138 139 140 141 142
		}

		/* Now parse data that is completely in this buffer */
		dma_q->is_partial_line = 0;

		while (bytes_parsed < buffer_size) {
			u32 bytes_used = 0;

143 144 145 146
			sav_eav = cx231xx_find_next_SAV_EAV(
				p_buffer + bytes_parsed,	/* p_buffer */
				buffer_size - bytes_parsed, /* buffer size */
				&bytes_used);	/* bytes used to get SAV/EAV */
147 148 149 150 151

			bytes_parsed += bytes_used;

			sav_eav &= 0xF0;
			if (sav_eav && (bytes_parsed < buffer_size)) {
152
				bytes_parsed += cx231xx_get_vbi_line(dev,
153 154 155
					dma_q, sav_eav,	/* SAV/EAV */
					p_buffer+bytes_parsed, /* p_buffer */
					buffer_size-bytes_parsed);/*buf size*/
156 157 158
			}
		}

159 160
		/* Save the last four bytes of the buffer so we can
		check the buffer boundary condition next time */
161 162
		memcpy(dma_q->partial_buf, p_buffer + buffer_size - 4, 4);
		bytes_parsed = 0;
163 164 165 166 167 168 169 170 171 172
	}

	return rc;
}

/* ------------------------------------------------------------------
	Vbi buf operations
   ------------------------------------------------------------------*/

static int
173 174
vbi_buffer_setup(struct videobuf_queue *vq, unsigned int *count,
		 unsigned int *size)
175 176
{
	struct cx231xx_fh *fh = vq->priv_data;
177 178
	struct cx231xx *dev = fh->dev;
	u32 height = 0;
179 180

	height = ((dev->norm & V4L2_STD_625_50) ?
181
		  PAL_VBI_LINES : NTSC_VBI_LINES);
182

183
	*size = (dev->width * height * 2 * 2);
184 185 186 187 188 189 190 191 192 193 194 195
	if (0 == *count)
		*count = CX231XX_DEF_VBI_BUF;

	if (*count < CX231XX_MIN_BUF)
		*count = CX231XX_MIN_BUF;

	return 0;
}

/* This is called *without* dev->slock held; please keep it that way */
static void free_buffer(struct videobuf_queue *vq, struct cx231xx_buffer *buf)
{
196 197
	struct cx231xx_fh *fh = vq->priv_data;
	struct cx231xx *dev = fh->dev;
198 199 200 201 202 203 204 205 206 207 208 209
	unsigned long flags = 0;
	if (in_interrupt())
		BUG();

	/* We used to wait for the buffer to finish here, but this didn't work
	   because, as we were keeping the state as VIDEOBUF_QUEUED,
	   videobuf_queue_cancel marked it as finished for us.
	   (Also, it could wedge forever if the hardware was misconfigured.)

	   This should be safe; by the time we get here, the buffer isn't
	   queued anymore. If we ever start marking the buffers as
	   VIDEOBUF_ACTIVE, it won't be, though.
210
	 */
211
	spin_lock_irqsave(&dev->vbi_mode.slock, flags);
212 213
	if (dev->vbi_mode.bulk_ctl.buf == buf)
		dev->vbi_mode.bulk_ctl.buf = NULL;
214 215 216 217 218 219 220 221
	spin_unlock_irqrestore(&dev->vbi_mode.slock, flags);

	videobuf_vmalloc_free(&buf->vb);
	buf->vb.state = VIDEOBUF_NEEDS_INIT;
}

static int
vbi_buffer_prepare(struct videobuf_queue *vq, struct videobuf_buffer *vb,
222
		   enum v4l2_field field)
223
{
224 225 226 227 228 229
	struct cx231xx_fh *fh = vq->priv_data;
	struct cx231xx_buffer *buf =
	    container_of(vb, struct cx231xx_buffer, vb);
	struct cx231xx *dev = fh->dev;
	int rc = 0, urb_init = 0;
	u32 height = 0;
230 231

	height = ((dev->norm & V4L2_STD_625_50) ?
232
		  PAL_VBI_LINES : NTSC_VBI_LINES);
233
	buf->vb.size = ((dev->width << 1) * height * 2);
234

235
	if (0 != buf->vb.baddr && buf->vb.bsize < buf->vb.size)
236 237
		return -EINVAL;

238
	buf->vb.width = dev->width;
239
	buf->vb.height = height;
240 241
	buf->vb.field = field;
	buf->vb.field = V4L2_FIELD_SEQ_TB;
242 243 244 245 246 247 248

	if (VIDEOBUF_NEEDS_INIT == buf->vb.state) {
		rc = videobuf_iolock(vq, &buf->vb, NULL);
		if (rc < 0)
			goto fail;
	}

249
	if (!dev->vbi_mode.bulk_ctl.num_bufs)
250 251 252 253
		urb_init = 1;

	if (urb_init) {
		rc = cx231xx_init_vbi_isoc(dev, CX231XX_NUM_VBI_PACKETS,
254 255 256
					   CX231XX_NUM_VBI_BUFS,
					   dev->vbi_mode.alt_max_pkt_size[0],
					   cx231xx_isoc_vbi_copy);
257 258 259 260 261 262 263
		if (rc < 0)
			goto fail;
	}

	buf->vb.state = VIDEOBUF_PREPARED;
	return 0;

264
fail:
265 266 267 268 269 270 271
	free_buffer(vq, buf);
	return rc;
}

static void
vbi_buffer_queue(struct videobuf_queue *vq, struct videobuf_buffer *vb)
{
272 273 274 275 276
	struct cx231xx_buffer *buf =
	    container_of(vb, struct cx231xx_buffer, vb);
	struct cx231xx_fh *fh = vq->priv_data;
	struct cx231xx *dev = fh->dev;
	struct cx231xx_dmaqueue *vidq = &dev->vbi_mode.vidq;
277 278 279 280 281 282 283

	buf->vb.state = VIDEOBUF_QUEUED;
	list_add_tail(&buf->vb.queue, &vidq->active);

}

static void vbi_buffer_release(struct videobuf_queue *vq,
284
			       struct videobuf_buffer *vb)
285
{
286 287
	struct cx231xx_buffer *buf =
	    container_of(vb, struct cx231xx_buffer, vb);
288 289 290 291 292 293


	free_buffer(vq, buf);
}

struct videobuf_queue_ops cx231xx_vbi_qops = {
294
	.buf_setup   = vbi_buffer_setup,
295
	.buf_prepare = vbi_buffer_prepare,
296
	.buf_queue   = vbi_buffer_queue,
297
	.buf_release = vbi_buffer_release,
298 299 300 301 302 303 304 305 306 307 308
};

/* ------------------------------------------------------------------
	URB control
   ------------------------------------------------------------------*/

/*
 * IRQ callback, called by URB callback
 */
static void cx231xx_irq_vbi_callback(struct urb *urb)
{
309 310 311 312
	struct cx231xx_dmaqueue *dma_q = urb->context;
	struct cx231xx_video_mode *vmode =
	    container_of(dma_q, struct cx231xx_video_mode, vidq);
	struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode);
313 314
	int rc;

315 316 317 318 319 320 321 322 323 324 325 326
	switch (urb->status) {
	case 0:		/* success */
	case -ETIMEDOUT:	/* NAK */
		break;
	case -ECONNRESET:	/* kill */
	case -ENOENT:
	case -ESHUTDOWN:
		return;
	default:		/* error */
		cx231xx_err(DRIVER_NAME "urb completition error %d.\n",
			    urb->status);
		break;
327 328 329 330
	}

	/* Copy data from URB */
	spin_lock(&dev->vbi_mode.slock);
331
	rc = dev->vbi_mode.bulk_ctl.bulk_copy(dev, urb);
332 333 334 335 336 337 338 339
	spin_unlock(&dev->vbi_mode.slock);

	/* Reset status */
	urb->status = 0;

	urb->status = usb_submit_urb(urb, GFP_ATOMIC);
	if (urb->status) {
		cx231xx_err(DRIVER_NAME "urb resubmit failed (error=%i)\n",
340
			    urb->status);
341 342 343 344 345 346 347 348 349 350 351
	}
}

/*
 * Stop and Deallocate URBs
 */
void cx231xx_uninit_vbi_isoc(struct cx231xx *dev)
{
	struct urb *urb;
	int i;

352
	cx231xx_info(DRIVER_NAME "cx231xx: called cx231xx_uninit_vbi_isoc\n");
353

354 355 356
	dev->vbi_mode.bulk_ctl.nfields = -1;
	for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
		urb = dev->vbi_mode.bulk_ctl.urb[i];
357
		if (urb) {
358 359 360 361
			if (!irqs_disabled())
				usb_kill_urb(urb);
			else
				usb_unlink_urb(urb);
362

363
			if (dev->vbi_mode.bulk_ctl.transfer_buffer[i]) {
364

365
				kfree(dev->vbi_mode.bulk_ctl.
366
				      transfer_buffer[i]);
367
				dev->vbi_mode.bulk_ctl.transfer_buffer[i] =
368
				    NULL;
369 370
			}
			usb_free_urb(urb);
371
			dev->vbi_mode.bulk_ctl.urb[i] = NULL;
372
		}
373
		dev->vbi_mode.bulk_ctl.transfer_buffer[i] = NULL;
374 375
	}

376 377
	kfree(dev->vbi_mode.bulk_ctl.urb);
	kfree(dev->vbi_mode.bulk_ctl.transfer_buffer);
378

379 380 381
	dev->vbi_mode.bulk_ctl.urb = NULL;
	dev->vbi_mode.bulk_ctl.transfer_buffer = NULL;
	dev->vbi_mode.bulk_ctl.num_bufs = 0;
382 383 384 385 386 387 388 389 390

	cx231xx_capture_start(dev, 0, Vbi);
}
EXPORT_SYMBOL_GPL(cx231xx_uninit_vbi_isoc);

/*
 * Allocate URBs and start IRQ
 */
int cx231xx_init_vbi_isoc(struct cx231xx *dev, int max_packets,
391
			  int num_bufs, int max_pkt_size,
392
			  int (*bulk_copy) (struct cx231xx *dev,
393
					    struct urb *urb))
394 395 396 397 398 399 400 401 402 403 404 405
{
	struct cx231xx_dmaqueue *dma_q = &dev->vbi_mode.vidq;
	int i;
	int sb_size, pipe;
	struct urb *urb;
	int rc;

	cx231xx_info(DRIVER_NAME "cx231xx: called cx231xx_prepare_isoc\n");

	/* De-allocates all pending stuff */
	cx231xx_uninit_vbi_isoc(dev);

406 407 408 409
	/* clear if any halt */
	usb_clear_halt(dev->udev,
		       usb_rcvbulkpipe(dev->udev,
				       dev->vbi_mode.end_point_addr));
410

411 412
	dev->vbi_mode.bulk_ctl.bulk_copy = bulk_copy;
	dev->vbi_mode.bulk_ctl.num_bufs = num_bufs;
413 414 415 416 417 418 419 420 421 422 423
	dma_q->pos = 0;
	dma_q->is_partial_line = 0;
	dma_q->last_sav = 0;
	dma_q->current_field = -1;
	dma_q->bytes_left_in_line = dev->width << 1;
	dma_q->lines_per_field = ((dev->norm & V4L2_STD_625_50) ?
				  PAL_VBI_LINES : NTSC_VBI_LINES);
	dma_q->lines_completed = 0;
	for (i = 0; i < 8; i++)
		dma_q->partial_buf[i] = 0;

424
	dev->vbi_mode.bulk_ctl.urb = kzalloc(sizeof(void *) * num_bufs,
425
					     GFP_KERNEL);
426
	if (!dev->vbi_mode.bulk_ctl.urb) {
427 428 429 430
		cx231xx_errdev("cannot alloc memory for usb buffers\n");
		return -ENOMEM;
	}

431
	dev->vbi_mode.bulk_ctl.transfer_buffer =
432
	    kzalloc(sizeof(void *) * num_bufs, GFP_KERNEL);
433
	if (!dev->vbi_mode.bulk_ctl.transfer_buffer) {
434
		cx231xx_errdev("cannot allocate memory for usbtransfer\n");
435
		kfree(dev->vbi_mode.bulk_ctl.urb);
436 437 438
		return -ENOMEM;
	}

439 440
	dev->vbi_mode.bulk_ctl.max_pkt_size = max_pkt_size;
	dev->vbi_mode.bulk_ctl.buf = NULL;
441

442
	sb_size = max_packets * dev->vbi_mode.bulk_ctl.max_pkt_size;
443 444

	/* allocate urbs and transfer buffers */
445
	for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
446 447 448

		urb = usb_alloc_urb(0, GFP_KERNEL);
		if (!urb) {
449
			cx231xx_err(DRIVER_NAME
450
				    ": cannot alloc bulk_ctl.urb %i\n", i);
451 452 453
			cx231xx_uninit_vbi_isoc(dev);
			return -ENOMEM;
		}
454
		dev->vbi_mode.bulk_ctl.urb[i] = urb;
455
		urb->transfer_flags = 0;
456

457
		dev->vbi_mode.bulk_ctl.transfer_buffer[i] =
458
		    kzalloc(sb_size, GFP_KERNEL);
459
		if (!dev->vbi_mode.bulk_ctl.transfer_buffer[i]) {
460 461 462
			cx231xx_err(DRIVER_NAME
				    ": unable to allocate %i bytes for transfer"
				    " buffer %i%s\n", sb_size, i,
463
				    in_interrupt() ? " while in int" : "");
464 465 466 467 468
			cx231xx_uninit_vbi_isoc(dev);
			return -ENOMEM;
		}

		pipe = usb_rcvbulkpipe(dev->udev, dev->vbi_mode.end_point_addr);
469
		usb_fill_bulk_urb(urb, dev->udev, pipe,
470
				  dev->vbi_mode.bulk_ctl.transfer_buffer[i],
471
				  sb_size, cx231xx_irq_vbi_callback, dma_q);
472 473 474 475 476
	}

	init_waitqueue_head(&dma_q->wq);

	/* submit urbs and enables IRQ */
477 478
	for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
		rc = usb_submit_urb(dev->vbi_mode.bulk_ctl.urb[i], GFP_ATOMIC);
479
		if (rc) {
480 481 482
			cx231xx_err(DRIVER_NAME
				    ": submit of urb %i failed (error=%i)\n", i,
				    rc);
483 484 485 486 487
			cx231xx_uninit_vbi_isoc(dev);
			return rc;
		}
	}

488
	cx231xx_capture_start(dev, 1, Vbi);
489 490 491

	return 0;
}
492
EXPORT_SYMBOL_GPL(cx231xx_init_vbi_isoc);
493

494 495
u32 cx231xx_get_vbi_line(struct cx231xx *dev, struct cx231xx_dmaqueue *dma_q,
			 u8 sav_eav, u8 *p_buffer, u32 buffer_size)
496
{
497 498
	u32 bytes_copied = 0;
	int current_field = -1;
499

500
	switch (sav_eav) {
501

502 503 504
	case SAV_VBI_FIELD1:
		current_field = 1;
		break;
505

506 507 508 509 510 511
	case SAV_VBI_FIELD2:
		current_field = 2;
		break;
	default:
		break;
	}
512

513 514
	if (current_field < 0)
		return bytes_copied;
515

516
	dma_q->last_sav = sav_eav;
517

518 519 520
	bytes_copied =
	    cx231xx_copy_vbi_line(dev, dma_q, p_buffer, buffer_size,
				  current_field);
521

522
	return bytes_copied;
523 524 525 526 527 528
}

/*
 * Announces that a buffer were filled and request the next
 */
static inline void vbi_buffer_filled(struct cx231xx *dev,
529 530
				     struct cx231xx_dmaqueue *dma_q,
				     struct cx231xx_buffer *buf)
531 532 533 534 535 536 537 538
{
	/* Advice that buffer was filled */
	/* cx231xx_info(DRIVER_NAME "[%p/%d] wakeup\n", buf, buf->vb.i); */

	buf->vb.state = VIDEOBUF_DONE;
	buf->vb.field_count++;
	do_gettimeofday(&buf->vb.ts);

539
	dev->vbi_mode.bulk_ctl.buf = NULL;
540 541 542 543 544

	list_del(&buf->vb.queue);
	wake_up(&buf->vb.done);
}

545
u32 cx231xx_copy_vbi_line(struct cx231xx *dev, struct cx231xx_dmaqueue *dma_q,
546
			  u8 *p_line, u32 length, int field_number)
547
{
548 549 550
	u32 bytes_to_copy;
	struct cx231xx_buffer *buf;
	u32 _line_size = dev->width * 2;
551

552 553
	if (dma_q->current_field == -1) {
		/* Just starting up */
554
		cx231xx_reset_vbi_buffer(dev, dma_q);
555 556 557 558
	}

	if (dma_q->current_field != field_number)
		dma_q->lines_completed = 0;
559

560
	/* get the buffer pointer */
561
	buf = dev->vbi_mode.bulk_ctl.buf;
562

563 564
	/* Remember the field number for next time */
	dma_q->current_field = field_number;
565

566 567 568
	bytes_to_copy = dma_q->bytes_left_in_line;
	if (bytes_to_copy > length)
		bytes_to_copy = length;
569

570 571 572 573 574 575
	if (dma_q->lines_completed >= dma_q->lines_per_field) {
		dma_q->bytes_left_in_line -= bytes_to_copy;
		dma_q->is_partial_line =
		    (dma_q->bytes_left_in_line == 0) ? 0 : 1;
		return 0;
	}
576

577
	dma_q->is_partial_line = 1;
578

579 580 581 582 583 584 585 586
	/* If we don't have a buffer, just return the number of bytes we would
	   have copied if we had a buffer. */
	if (!buf) {
		dma_q->bytes_left_in_line -= bytes_to_copy;
		dma_q->is_partial_line =
		    (dma_q->bytes_left_in_line == 0) ? 0 : 1;
		return bytes_to_copy;
	}
587

588 589
	/* copy the data to video buffer */
	cx231xx_do_vbi_copy(dev, dma_q, p_line, bytes_to_copy);
590

591 592
	dma_q->pos += bytes_to_copy;
	dma_q->bytes_left_in_line -= bytes_to_copy;
593

594
	if (dma_q->bytes_left_in_line == 0) {
595

596 597 598
		dma_q->bytes_left_in_line = _line_size;
		dma_q->lines_completed++;
		dma_q->is_partial_line = 0;
599

600
		if (cx231xx_is_vbi_buffer_done(dev, dma_q) && buf) {
601

602
			vbi_buffer_filled(dev, dma_q, buf);
603

604 605
			dma_q->pos = 0;
			dma_q->lines_completed = 0;
606
			cx231xx_reset_vbi_buffer(dev, dma_q);
607 608
		}
	}
609

610
	return bytes_to_copy;
611 612 613 614 615 616
}

/*
 * video-buf generic routine to get the next available buffer
 */
static inline void get_next_vbi_buf(struct cx231xx_dmaqueue *dma_q,
617
				    struct cx231xx_buffer **buf)
618
{
619 620 621
	struct cx231xx_video_mode *vmode =
	    container_of(dma_q, struct cx231xx_video_mode, vidq);
	struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode);
622 623 624
	char *outp;

	if (list_empty(&dma_q->active)) {
625
		cx231xx_err(DRIVER_NAME ": No active queue to serve\n");
626
		dev->vbi_mode.bulk_ctl.buf = NULL;
627 628 629 630 631 632 633
		*buf = NULL;
		return;
	}

	/* Get the next buffer */
	*buf = list_entry(dma_q->active.next, struct cx231xx_buffer, vb.queue);

L
Lucas De Marchi 已提交
634
	/* Cleans up buffer - Useful for testing for frame/URB loss */
635 636 637
	outp = videobuf_to_vmalloc(&(*buf)->vb);
	memset(outp, 0, (*buf)->vb.size);

638
	dev->vbi_mode.bulk_ctl.buf = *buf;
639 640 641 642

	return;
}

643 644
void cx231xx_reset_vbi_buffer(struct cx231xx *dev,
			      struct cx231xx_dmaqueue *dma_q)
645
{
646
	struct cx231xx_buffer *buf;
647

648
	buf = dev->vbi_mode.bulk_ctl.buf;
649

650 651 652
	if (buf == NULL) {
		/* first try to get the buffer */
		get_next_vbi_buf(dma_q, &buf);
653

654 655 656
		dma_q->pos = 0;
		dma_q->current_field = -1;
	}
657

658 659
	dma_q->bytes_left_in_line = dev->width << 1;
	dma_q->lines_completed = 0;
660 661 662
}

int cx231xx_do_vbi_copy(struct cx231xx *dev, struct cx231xx_dmaqueue *dma_q,
663
			u8 *p_buffer, u32 bytes_to_copy)
664
{
665 666 667 668 669 670
	u8 *p_out_buffer = NULL;
	u32 current_line_bytes_copied = 0;
	struct cx231xx_buffer *buf;
	u32 _line_size = dev->width << 1;
	void *startwrite;
	int offset, lencopy;
671

672
	buf = dev->vbi_mode.bulk_ctl.buf;
673

674 675
	if (buf == NULL)
		return -EINVAL;
676 677 678

	p_out_buffer = videobuf_to_vmalloc(&buf->vb);

679 680 681 682
	if (dma_q->bytes_left_in_line != _line_size) {
		current_line_bytes_copied =
		    _line_size - dma_q->bytes_left_in_line;
	}
683

684 685
	offset = (dma_q->lines_completed * _line_size) +
		 current_line_bytes_copied;
686

687 688 689 690 691
	if (dma_q->current_field == 2) {
		/* Populate the second half of the frame */
		offset += (dev->width * 2 * dma_q->lines_per_field);
	}

692
	/* prepare destination address */
693 694
	startwrite = p_out_buffer + offset;

695 696
	lencopy = dma_q->bytes_left_in_line > bytes_to_copy ?
		  bytes_to_copy : dma_q->bytes_left_in_line;
697

698
	memcpy(startwrite, p_buffer, lencopy);
699

700
	return 0;
701 702
}

703 704
u8 cx231xx_is_vbi_buffer_done(struct cx231xx *dev,
			      struct cx231xx_dmaqueue *dma_q)
705
{
706
	u32 height = 0;
707 708

	height = ((dev->norm & V4L2_STD_625_50) ?
709
		  PAL_VBI_LINES : NTSC_VBI_LINES);
710 711 712 713
	if (dma_q->lines_completed == height && dma_q->current_field == 2)
		return 1;
	else
		return 0;
714
}