xprtsock.c 26.6 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13
/*
 * linux/net/sunrpc/xprtsock.c
 *
 * Client-side transport implementation for sockets.
 *
 * TCP callback races fixes (C) 1998 Red Hat Software <alan@redhat.com>
 * TCP send fixes (C) 1998 Red Hat Software <alan@redhat.com>
 * TCP NFS related read + write fixes
 *  (C) 1999 Dave Airlie, University of Limerick, Ireland <airlied@linux.ie>
 *
 * Rewrite of larges part of the code in order to stabilize TCP stuff.
 * Fix behaviour when socket buffer is full.
 *  (C) 1999 Trond Myklebust <trond.myklebust@fys.uio.no>
14 15
 *
 * IP socket transport implementation, (C) 2005 Chuck Lever <cel@netapp.com>
16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37
 */

#include <linux/types.h>
#include <linux/slab.h>
#include <linux/capability.h>
#include <linux/sched.h>
#include <linux/pagemap.h>
#include <linux/errno.h>
#include <linux/socket.h>
#include <linux/in.h>
#include <linux/net.h>
#include <linux/mm.h>
#include <linux/udp.h>
#include <linux/tcp.h>
#include <linux/sunrpc/clnt.h>
#include <linux/file.h>

#include <net/sock.h>
#include <net/checksum.h>
#include <net/udp.h>
#include <net/tcp.h>

38 39 40 41 42
/*
 * Maximum port number to use when requesting a reserved port.
 */
#define XS_MAX_RESVPORT		(800U)

43 44
#ifdef RPC_DEBUG
# undef  RPC_DEBUG_DATA
45
# define RPCDBG_FACILITY	RPCDBG_TRANS
46 47 48
#endif

#ifdef RPC_DEBUG_DATA
49
static void xs_pktdump(char *msg, u32 *packet, unsigned int count)
50
{
51 52
	u8 *buf = (u8 *) packet;
	int j;
53 54 55 56 57 58 59 60 61 62 63 64 65 66

	dprintk("RPC:      %s\n", msg);
	for (j = 0; j < count && j < 128; j += 4) {
		if (!(j & 31)) {
			if (j)
				dprintk("\n");
			dprintk("0x%04x ", j);
		}
		dprintk("%02x%02x%02x%02x ",
			buf[j], buf[j+1], buf[j+2], buf[j+3]);
	}
	dprintk("\n");
}
#else
67
static inline void xs_pktdump(char *msg, u32 *packet, unsigned int count)
68 69 70 71 72
{
	/* NOP */
}
#endif

73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107
#define XS_SENDMSG_FLAGS	(MSG_DONTWAIT | MSG_NOSIGNAL)

static inline int xs_send_head(struct socket *sock, struct sockaddr *addr, int addrlen, struct xdr_buf *xdr, unsigned int base, unsigned int len)
{
	struct kvec iov = {
		.iov_base	= xdr->head[0].iov_base + base,
		.iov_len	= len - base,
	};
	struct msghdr msg = {
		.msg_name	= addr,
		.msg_namelen	= addrlen,
		.msg_flags	= XS_SENDMSG_FLAGS,
	};

	if (xdr->len > len)
		msg.msg_flags |= MSG_MORE;

	if (likely(iov.iov_len))
		return kernel_sendmsg(sock, &msg, &iov, 1, iov.iov_len);
	return kernel_sendmsg(sock, &msg, NULL, 0, 0);
}

static int xs_send_tail(struct socket *sock, struct xdr_buf *xdr, unsigned int base, unsigned int len)
{
	struct kvec iov = {
		.iov_base	= xdr->tail[0].iov_base + base,
		.iov_len	= len - base,
	};
	struct msghdr msg = {
		.msg_flags	= XS_SENDMSG_FLAGS,
	};

	return kernel_sendmsg(sock, &msg, &iov, 1, iov.iov_len);
}

108 109 110 111 112 113 114 115
/**
 * xs_sendpages - write pages directly to a socket
 * @sock: socket to send on
 * @addr: UDP only -- address of destination
 * @addrlen: UDP only -- length of destination address
 * @xdr: buffer containing this request
 * @base: starting position in the buffer
 *
116
 */
117
static int xs_sendpages(struct socket *sock, struct sockaddr *addr, int addrlen, struct xdr_buf *xdr, unsigned int base)
118 119 120 121 122 123 124 125
{
	struct page **ppage = xdr->pages;
	unsigned int len, pglen = xdr->page_len;
	int err, ret = 0;
	ssize_t (*sendpage)(struct socket *, struct page *, int, size_t, int);

	len = xdr->head[0].iov_len;
	if (base < len || (addr != NULL && base == 0)) {
126
		err = xs_send_head(sock, addr, addrlen, xdr, base, len);
127 128 129 130
		if (ret == 0)
			ret = err;
		else if (err > 0)
			ret += err;
131
		if (err != (len - base))
132 133 134 135 136
			goto out;
		base = 0;
	} else
		base -= len;

137
	if (unlikely(pglen == 0))
138
		goto copy_tail;
139
	if (unlikely(base >= pglen)) {
140 141 142 143 144
		base -= pglen;
		goto copy_tail;
	}
	if (base || xdr->page_base) {
		pglen -= base;
145
		base += xdr->page_base;
146 147 148 149 150 151
		ppage += base >> PAGE_CACHE_SHIFT;
		base &= ~PAGE_CACHE_MASK;
	}

	sendpage = sock->ops->sendpage ? : sock_no_sendpage;
	do {
152
		int flags = XS_SENDMSG_FLAGS;
153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178

		len = PAGE_CACHE_SIZE;
		if (base)
			len -= base;
		if (pglen < len)
			len = pglen;

		if (pglen != len || xdr->tail[0].iov_len != 0)
			flags |= MSG_MORE;

		/* Hmm... We might be dealing with highmem pages */
		if (PageHighMem(*ppage))
			sendpage = sock_no_sendpage;
		err = sendpage(sock, *ppage, base, len, flags);
		if (ret == 0)
			ret = err;
		else if (err > 0)
			ret += err;
		if (err != len)
			goto out;
		base = 0;
		ppage++;
	} while ((pglen -= len) != 0);
copy_tail:
	len = xdr->tail[0].iov_len;
	if (base < len) {
179
		err = xs_send_tail(sock, xdr, base, len);
180 181 182 183 184 185 186 187 188
		if (ret == 0)
			ret = err;
		else if (err > 0)
			ret += err;
	}
out:
	return ret;
}

189 190 191 192 193
/**
 * xs_sendmsg - write an RPC request to a socket
 * @xprt: generic transport
 * @req: the RPC request to write
 *
194
 */
195
static int xs_sendmsg(struct rpc_xprt *xprt, struct rpc_rqst *req)
196
{
197 198
	struct socket *sock = xprt->sock;
	struct xdr_buf *xdr = &req->rq_snd_buf;
199 200
	struct sockaddr *addr = NULL;
	int addrlen = 0;
201 202
	unsigned int skip;
	int result;
203 204 205 206

	if (!sock)
		return -ENOTCONN;

207
	xs_pktdump("packet data:",
208 209 210 211 212 213 214 215
				req->rq_svec->iov_base,
				req->rq_svec->iov_len);

	/* For UDP, we need to provide an address */
	if (!xprt->stream) {
		addr = (struct sockaddr *) &xprt->addr;
		addrlen = sizeof(xprt->addr);
	}
216
	/* Don't repeat bytes */
217 218 219
	skip = req->rq_bytes_sent;

	clear_bit(SOCK_ASYNC_NOSPACE, &sock->flags);
220
	result = xs_sendpages(sock, addr, addrlen, xdr, skip);
221

222
	dprintk("RPC:      xs_sendmsg(%d) = %d\n", xdr->len - skip, result);
223 224 225 226 227 228 229

	if (result >= 0)
		return result;

	switch (result) {
	case -ECONNREFUSED:
		/* When the server has died, an ICMP port unreachable message
230
		 * prompts ECONNREFUSED. */
231 232 233 234 235 236 237 238 239 240
	case -EAGAIN:
		break;
	case -ECONNRESET:
	case -ENOTCONN:
	case -EPIPE:
		/* connection broken */
		if (xprt->stream)
			result = -ENOTCONN;
		break;
	default:
241
		break;
242 243 244 245
	}
	return result;
}

246 247 248 249 250 251 252 253 254 255 256 257 258 259
/**
 * xs_send_request - write an RPC request to a socket
 * @task: address of RPC task that manages the state of an RPC request
 *
 * Return values:
 *      0:  The request has been sent
 * EAGAIN:  The socket was blocked, please call again later to
 *          complete the request
 *  other:  Some other error occured, the request was not sent
 *
 * XXX: In the case of soft timeouts, should we eventually give up
 *      if the socket is not able to make progress?
 */
static int xs_send_request(struct rpc_task *task)
260 261 262 263 264 265 266 267
{
	struct rpc_rqst *req = task->tk_rqstp;
	struct rpc_xprt *xprt = req->rq_xprt;
	int status, retry = 0;

	/* set up everything as needed. */
	/* Write the record marker */
	if (xprt->stream) {
268
		u32 *marker = req->rq_svec[0].iov_base;
269 270 271 272 273 274

		*marker = htonl(0x80000000|(req->rq_slen-sizeof(*marker)));
	}

	/* Continue transmitting the packet/record. We must be careful
	 * to cope with writespace callbacks arriving _after_ we have
275
	 * called sendmsg().
276 277 278
	 */
	while (1) {
		req->rq_xtime = jiffies;
279
		status = xs_sendmsg(xprt, req);
280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310

		if (status < 0)
			break;

		if (xprt->stream) {
			req->rq_bytes_sent += status;

			/* If we've sent the entire packet, immediately
			 * reset the count of bytes sent. */
			if (req->rq_bytes_sent >= req->rq_slen) {
				req->rq_bytes_sent = 0;
				return 0;
			}
		} else {
			if (status >= req->rq_slen)
				return 0;
			status = -EAGAIN;
			break;
		}

		dprintk("RPC: %4d xmit incomplete (%d left of %d)\n",
				task->tk_pid, req->rq_slen - req->rq_bytes_sent,
				req->rq_slen);

		status = -EAGAIN;
		if (retry++ > 50)
			break;
	}

	if (status == -EAGAIN) {
		if (test_bit(SOCK_ASYNC_NOSPACE, &xprt->sock->flags)) {
311
			/* Protect against races with write_space */
C
Chuck Lever 已提交
312
			spin_lock_bh(&xprt->transport_lock);
313 314 315
			/* Don't race with disconnect */
			if (!xprt_connected(xprt))
				task->tk_status = -ENOTCONN;
316 317
			else if (test_bit(SOCK_NOSPACE, &xprt->sock->flags))
				xprt_wait_for_buffer_space(task);
C
Chuck Lever 已提交
318
			spin_unlock_bh(&xprt->transport_lock);
319 320 321 322 323 324 325 326
			return status;
		}
		/* Keep holding the socket if it is blocked */
		rpc_delay(task, HZ>>4);
	}
	return status;
}

327 328 329 330
/**
 * xs_close - close a socket
 * @xprt: transport
 *
331
 */
332
static void xs_close(struct rpc_xprt *xprt)
333
{
334 335
	struct socket *sock = xprt->sock;
	struct sock *sk = xprt->inet;
336 337 338 339

	if (!sk)
		return;

340 341
	dprintk("RPC:      xs_close xprt %p\n", xprt);

342 343 344 345
	write_lock_bh(&sk->sk_callback_lock);
	xprt->inet = NULL;
	xprt->sock = NULL;

346 347
	sk->sk_user_data = NULL;
	sk->sk_data_ready = xprt->old_data_ready;
348
	sk->sk_state_change = xprt->old_state_change;
349
	sk->sk_write_space = xprt->old_write_space;
350 351
	write_unlock_bh(&sk->sk_callback_lock);

352
	sk->sk_no_check = 0;
353 354 355 356

	sock_release(sock);
}

357 358 359 360 361 362
/**
 * xs_destroy - prepare to shutdown a transport
 * @xprt: doomed transport
 *
 */
static void xs_destroy(struct rpc_xprt *xprt)
363
{
364 365
	dprintk("RPC:      xs_destroy xprt %p\n", xprt);

366
	cancel_delayed_work(&xprt->connect_worker);
367 368 369
	flush_scheduled_work();

	xprt_disconnect(xprt);
370
	xs_close(xprt);
371 372 373
	kfree(xprt->slot);
}

374 375 376 377 378 379 380 381 382 383
static inline struct rpc_xprt *xprt_from_sock(struct sock *sk)
{
	return (struct rpc_xprt *) sk->sk_user_data;
}

/**
 * xs_udp_data_ready - "data ready" callback for UDP sockets
 * @sk: socket with data to read
 * @len: how much data to read
 *
384
 */
385
static void xs_udp_data_ready(struct sock *sk, int len)
386
{
387 388
	struct rpc_task *task;
	struct rpc_xprt *xprt;
389
	struct rpc_rqst *rovr;
390
	struct sk_buff *skb;
391 392 393 394
	int err, repsize, copied;
	u32 _xid, *xp;

	read_lock(&sk->sk_callback_lock);
395 396
	dprintk("RPC:      xs_udp_data_ready...\n");
	if (!(xprt = xprt_from_sock(sk)))
397 398 399 400 401 402 403 404 405 406
		goto out;

	if ((skb = skb_recv_datagram(sk, 0, 1, &err)) == NULL)
		goto out;

	if (xprt->shutdown)
		goto dropit;

	repsize = skb->len - sizeof(struct udphdr);
	if (repsize < 4) {
407
		dprintk("RPC:      impossible RPC reply size %d!\n", repsize);
408 409 410 411 412 413 414 415 416 417
		goto dropit;
	}

	/* Copy the XID from the skb... */
	xp = skb_header_pointer(skb, sizeof(struct udphdr),
				sizeof(_xid), &_xid);
	if (xp == NULL)
		goto dropit;

	/* Look up and lock the request corresponding to the given XID */
C
Chuck Lever 已提交
418
	spin_lock(&xprt->transport_lock);
419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438
	rovr = xprt_lookup_rqst(xprt, *xp);
	if (!rovr)
		goto out_unlock;
	task = rovr->rq_task;

	dprintk("RPC: %4d received reply\n", task->tk_pid);

	if ((copied = rovr->rq_private_buf.buflen) > repsize)
		copied = repsize;

	/* Suck it into the iovec, verify checksum if not done by hw. */
	if (csum_partial_copy_to_xdr(&rovr->rq_private_buf, skb))
		goto out_unlock;

	/* Something worked... */
	dst_confirm(skb->dst);

	xprt_complete_rqst(xprt, rovr, copied);

 out_unlock:
C
Chuck Lever 已提交
439
	spin_unlock(&xprt->transport_lock);
440 441 442 443 444 445
 dropit:
	skb_free_datagram(sk, skb);
 out:
	read_unlock(&sk->sk_callback_lock);
}

446
static inline size_t xs_tcp_copy_data(skb_reader_t *desc, void *p, size_t len)
447 448 449 450 451 452 453 454 455 456 457 458 459 460 461
{
	if (len > desc->count)
		len = desc->count;
	if (skb_copy_bits(desc->skb, desc->offset, p, len)) {
		dprintk("RPC:      failed to copy %zu bytes from skb. %zu bytes remain\n",
				len, desc->count);
		return 0;
	}
	desc->offset += len;
	desc->count -= len;
	dprintk("RPC:      copied %zu bytes from skb. %zu bytes remain\n",
			len, desc->count);
	return len;
}

462
static inline void xs_tcp_read_fraghdr(struct rpc_xprt *xprt, skb_reader_t *desc)
463 464 465 466 467 468
{
	size_t len, used;
	char *p;

	p = ((char *) &xprt->tcp_recm) + xprt->tcp_offset;
	len = sizeof(xprt->tcp_recm) - xprt->tcp_offset;
469
	used = xs_tcp_copy_data(desc, p, len);
470 471 472 473 474 475 476 477 478 479 480 481 482
	xprt->tcp_offset += used;
	if (used != len)
		return;
	xprt->tcp_reclen = ntohl(xprt->tcp_recm);
	if (xprt->tcp_reclen & 0x80000000)
		xprt->tcp_flags |= XPRT_LAST_FRAG;
	else
		xprt->tcp_flags &= ~XPRT_LAST_FRAG;
	xprt->tcp_reclen &= 0x7fffffff;
	xprt->tcp_flags &= ~XPRT_COPY_RECM;
	xprt->tcp_offset = 0;
	/* Sanity check of the record length */
	if (xprt->tcp_reclen < 4) {
483
		dprintk("RPC:      invalid TCP record fragment length\n");
484
		xprt_disconnect(xprt);
485
		return;
486 487 488 489 490
	}
	dprintk("RPC:      reading TCP record fragment of length %d\n",
			xprt->tcp_reclen);
}

491
static void xs_tcp_check_recm(struct rpc_xprt *xprt)
492 493 494 495 496 497 498 499 500 501 502 503 504 505
{
	dprintk("RPC:      xprt = %p, tcp_copied = %lu, tcp_offset = %u, tcp_reclen = %u, tcp_flags = %lx\n",
			xprt, xprt->tcp_copied, xprt->tcp_offset, xprt->tcp_reclen, xprt->tcp_flags);
	if (xprt->tcp_offset == xprt->tcp_reclen) {
		xprt->tcp_flags |= XPRT_COPY_RECM;
		xprt->tcp_offset = 0;
		if (xprt->tcp_flags & XPRT_LAST_FRAG) {
			xprt->tcp_flags &= ~XPRT_COPY_DATA;
			xprt->tcp_flags |= XPRT_COPY_XID;
			xprt->tcp_copied = 0;
		}
	}
}

506
static inline void xs_tcp_read_xid(struct rpc_xprt *xprt, skb_reader_t *desc)
507 508 509 510 511 512 513
{
	size_t len, used;
	char *p;

	len = sizeof(xprt->tcp_xid) - xprt->tcp_offset;
	dprintk("RPC:      reading XID (%Zu bytes)\n", len);
	p = ((char *) &xprt->tcp_xid) + xprt->tcp_offset;
514
	used = xs_tcp_copy_data(desc, p, len);
515 516 517 518 519 520 521 522
	xprt->tcp_offset += used;
	if (used != len)
		return;
	xprt->tcp_flags &= ~XPRT_COPY_XID;
	xprt->tcp_flags |= XPRT_COPY_DATA;
	xprt->tcp_copied = 4;
	dprintk("RPC:      reading reply for XID %08x\n",
						ntohl(xprt->tcp_xid));
523
	xs_tcp_check_recm(xprt);
524 525
}

526
static inline void xs_tcp_read_request(struct rpc_xprt *xprt, skb_reader_t *desc)
527 528 529 530 531 532 533
{
	struct rpc_rqst *req;
	struct xdr_buf *rcvbuf;
	size_t len;
	ssize_t r;

	/* Find and lock the request corresponding to this xid */
C
Chuck Lever 已提交
534
	spin_lock(&xprt->transport_lock);
535 536 537 538 539
	req = xprt_lookup_rqst(xprt, xprt->tcp_xid);
	if (!req) {
		xprt->tcp_flags &= ~XPRT_COPY_DATA;
		dprintk("RPC:      XID %08x request not found!\n",
				ntohl(xprt->tcp_xid));
C
Chuck Lever 已提交
540
		spin_unlock(&xprt->transport_lock);
541 542 543 544 545 546 547 548 549 550 551 552
		return;
	}

	rcvbuf = &req->rq_private_buf;
	len = desc->count;
	if (len > xprt->tcp_reclen - xprt->tcp_offset) {
		skb_reader_t my_desc;

		len = xprt->tcp_reclen - xprt->tcp_offset;
		memcpy(&my_desc, desc, sizeof(my_desc));
		my_desc.count = len;
		r = xdr_partial_copy_from_skb(rcvbuf, xprt->tcp_copied,
553
					  &my_desc, xs_tcp_copy_data);
554 555 556 557
		desc->count -= r;
		desc->offset += r;
	} else
		r = xdr_partial_copy_from_skb(rcvbuf, xprt->tcp_copied,
558
					  desc, xs_tcp_copy_data);
559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599

	if (r > 0) {
		xprt->tcp_copied += r;
		xprt->tcp_offset += r;
	}
	if (r != len) {
		/* Error when copying to the receive buffer,
		 * usually because we weren't able to allocate
		 * additional buffer pages. All we can do now
		 * is turn off XPRT_COPY_DATA, so the request
		 * will not receive any additional updates,
		 * and time out.
		 * Any remaining data from this record will
		 * be discarded.
		 */
		xprt->tcp_flags &= ~XPRT_COPY_DATA;
		dprintk("RPC:      XID %08x truncated request\n",
				ntohl(xprt->tcp_xid));
		dprintk("RPC:      xprt = %p, tcp_copied = %lu, tcp_offset = %u, tcp_reclen = %u\n",
				xprt, xprt->tcp_copied, xprt->tcp_offset, xprt->tcp_reclen);
		goto out;
	}

	dprintk("RPC:      XID %08x read %Zd bytes\n",
			ntohl(xprt->tcp_xid), r);
	dprintk("RPC:      xprt = %p, tcp_copied = %lu, tcp_offset = %u, tcp_reclen = %u\n",
			xprt, xprt->tcp_copied, xprt->tcp_offset, xprt->tcp_reclen);

	if (xprt->tcp_copied == req->rq_private_buf.buflen)
		xprt->tcp_flags &= ~XPRT_COPY_DATA;
	else if (xprt->tcp_offset == xprt->tcp_reclen) {
		if (xprt->tcp_flags & XPRT_LAST_FRAG)
			xprt->tcp_flags &= ~XPRT_COPY_DATA;
	}

out:
	if (!(xprt->tcp_flags & XPRT_COPY_DATA)) {
		dprintk("RPC: %4d received reply complete\n",
				req->rq_task->tk_pid);
		xprt_complete_rqst(xprt, req, xprt->tcp_copied);
	}
C
Chuck Lever 已提交
600
	spin_unlock(&xprt->transport_lock);
601
	xs_tcp_check_recm(xprt);
602 603
}

604
static inline void xs_tcp_read_discard(struct rpc_xprt *xprt, skb_reader_t *desc)
605 606 607 608 609 610 611 612 613 614
{
	size_t len;

	len = xprt->tcp_reclen - xprt->tcp_offset;
	if (len > desc->count)
		len = desc->count;
	desc->count -= len;
	desc->offset += len;
	xprt->tcp_offset += len;
	dprintk("RPC:      discarded %Zu bytes\n", len);
615
	xs_tcp_check_recm(xprt);
616 617
}

618
static int xs_tcp_data_recv(read_descriptor_t *rd_desc, struct sk_buff *skb, unsigned int offset, size_t len)
619 620 621 622 623 624 625
{
	struct rpc_xprt *xprt = rd_desc->arg.data;
	skb_reader_t desc = {
		.skb	= skb,
		.offset	= offset,
		.count	= len,
		.csum	= 0
626
	};
627

628
	dprintk("RPC:      xs_tcp_data_recv started\n");
629 630 631 632
	do {
		/* Read in a new fragment marker if necessary */
		/* Can we ever really expect to get completely empty fragments? */
		if (xprt->tcp_flags & XPRT_COPY_RECM) {
633
			xs_tcp_read_fraghdr(xprt, &desc);
634 635 636 637
			continue;
		}
		/* Read in the xid if necessary */
		if (xprt->tcp_flags & XPRT_COPY_XID) {
638
			xs_tcp_read_xid(xprt, &desc);
639 640 641 642
			continue;
		}
		/* Read in the request data */
		if (xprt->tcp_flags & XPRT_COPY_DATA) {
643
			xs_tcp_read_request(xprt, &desc);
644 645 646
			continue;
		}
		/* Skip over any trailing bytes on short reads */
647
		xs_tcp_read_discard(xprt, &desc);
648
	} while (desc.count);
649
	dprintk("RPC:      xs_tcp_data_recv done\n");
650 651 652
	return len - desc.count;
}

653 654 655 656 657 658 659
/**
 * xs_tcp_data_ready - "data ready" callback for TCP sockets
 * @sk: socket with data to read
 * @bytes: how much data to read
 *
 */
static void xs_tcp_data_ready(struct sock *sk, int bytes)
660 661 662 663 664
{
	struct rpc_xprt *xprt;
	read_descriptor_t rd_desc;

	read_lock(&sk->sk_callback_lock);
665 666
	dprintk("RPC:      xs_tcp_data_ready...\n");
	if (!(xprt = xprt_from_sock(sk)))
667 668 669 670
		goto out;
	if (xprt->shutdown)
		goto out;

671
	/* We use rd_desc to pass struct xprt to xs_tcp_data_recv */
672 673
	rd_desc.arg.data = xprt;
	rd_desc.count = 65536;
674
	tcp_read_sock(sk, &rd_desc, xs_tcp_data_recv);
675 676 677 678
out:
	read_unlock(&sk->sk_callback_lock);
}

679 680 681 682 683 684
/**
 * xs_tcp_state_change - callback to handle TCP socket state changes
 * @sk: socket whose state has changed
 *
 */
static void xs_tcp_state_change(struct sock *sk)
685
{
686
	struct rpc_xprt *xprt;
687 688 689 690

	read_lock(&sk->sk_callback_lock);
	if (!(xprt = xprt_from_sock(sk)))
		goto out;
691
	dprintk("RPC:      xs_tcp_state_change client %p...\n", xprt);
692 693 694 695 696 697 698
	dprintk("RPC:      state %x conn %d dead %d zapped %d\n",
				sk->sk_state, xprt_connected(xprt),
				sock_flag(sk, SOCK_DEAD),
				sock_flag(sk, SOCK_ZAPPED));

	switch (sk->sk_state) {
	case TCP_ESTABLISHED:
C
Chuck Lever 已提交
699
		spin_lock_bh(&xprt->transport_lock);
700 701 702 703 704 705
		if (!xprt_test_and_set_connected(xprt)) {
			/* Reset TCP record info */
			xprt->tcp_offset = 0;
			xprt->tcp_reclen = 0;
			xprt->tcp_copied = 0;
			xprt->tcp_flags = XPRT_COPY_RECM | XPRT_COPY_XID;
706
			xprt_wake_pending_tasks(xprt, 0);
707
		}
C
Chuck Lever 已提交
708
		spin_unlock_bh(&xprt->transport_lock);
709 710 711 712 713 714 715 716 717 718 719 720
		break;
	case TCP_SYN_SENT:
	case TCP_SYN_RECV:
		break;
	default:
		xprt_disconnect(xprt);
		break;
	}
 out:
	read_unlock(&sk->sk_callback_lock);
}

721
/**
722 723
 * xs_udp_write_space - callback invoked when socket buffer space
 *                             becomes available
724 725
 * @sk: socket whose state has changed
 *
726 727
 * Called when more output buffer space is available for this socket.
 * We try not to wake our writers until they can make "significant"
728
 * progress, otherwise we'll waste resources thrashing kernel_sendmsg
729 730
 * with a bunch of small requests.
 */
731
static void xs_udp_write_space(struct sock *sk)
732 733 734
{
	read_lock(&sk->sk_callback_lock);

735 736 737 738 739 740
	/* from net/core/sock.c:sock_def_write_space */
	if (sock_writeable(sk)) {
		struct socket *sock;
		struct rpc_xprt *xprt;

		if (unlikely(!(sock = sk->sk_socket)))
741
			goto out;
742 743 744
		if (unlikely(!(xprt = xprt_from_sock(sk))))
			goto out;
		if (unlikely(!test_and_clear_bit(SOCK_NOSPACE, &sock->flags)))
745
			goto out;
746 747

		xprt_write_space(xprt);
748 749
	}

750 751 752
 out:
	read_unlock(&sk->sk_callback_lock);
}
753

754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783
/**
 * xs_tcp_write_space - callback invoked when socket buffer space
 *                             becomes available
 * @sk: socket whose state has changed
 *
 * Called when more output buffer space is available for this socket.
 * We try not to wake our writers until they can make "significant"
 * progress, otherwise we'll waste resources thrashing kernel_sendmsg
 * with a bunch of small requests.
 */
static void xs_tcp_write_space(struct sock *sk)
{
	read_lock(&sk->sk_callback_lock);

	/* from net/core/stream.c:sk_stream_write_space */
	if (sk_stream_wspace(sk) >= sk_stream_min_wspace(sk)) {
		struct socket *sock;
		struct rpc_xprt *xprt;

		if (unlikely(!(sock = sk->sk_socket)))
			goto out;
		if (unlikely(!(xprt = xprt_from_sock(sk))))
			goto out;
		if (unlikely(!test_and_clear_bit(SOCK_NOSPACE, &sock->flags)))
			goto out;

		xprt_write_space(xprt);
	}

 out:
784 785 786
	read_unlock(&sk->sk_callback_lock);
}

787 788 789 790 791 792 793
/**
 * xs_set_buffer_size - set send and receive limits
 * @xprt: generic transport
 *
 * Set socket send and receive limits based on the
 * sndsize and rcvsize fields in the generic transport
 * structure. This applies only to UDP sockets.
794
 */
795
static void xs_set_buffer_size(struct rpc_xprt *xprt)
796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811
{
	struct sock *sk = xprt->inet;

	if (xprt->stream)
		return;
	if (xprt->rcvsize) {
		sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
		sk->sk_rcvbuf = xprt->rcvsize * xprt->max_reqs *  2;
	}
	if (xprt->sndsize) {
		sk->sk_userlocks |= SOCK_SNDBUF_LOCK;
		sk->sk_sndbuf = xprt->sndsize * xprt->max_reqs * 2;
		sk->sk_write_space(sk);
	}
}

812
static int xs_bindresvport(struct rpc_xprt *xprt, struct socket *sock)
813 814 815 816
{
	struct sockaddr_in myaddr = {
		.sin_family = AF_INET,
	};
817
	int err, port;
818 819 820 821 822 823 824 825 826

	/* Were we already bound to a given port? Try to reuse it */
	port = xprt->port;
	do {
		myaddr.sin_port = htons(port);
		err = sock->ops->bind(sock, (struct sockaddr *) &myaddr,
						sizeof(myaddr));
		if (err == 0) {
			xprt->port = port;
827 828
			dprintk("RPC:      xs_bindresvport bound to port %u\n",
					port);
829 830 831
			return 0;
		}
		if (--port == 0)
832
			port = XS_MAX_RESVPORT;
833 834
	} while (err == -EADDRINUSE && port != xprt->port);

835
	dprintk("RPC:      can't bind to reserved port (%d).\n", -err);
836 837 838
	return err;
}

839 840 841 842 843 844 845
/**
 * xs_udp_connect_worker - set up a UDP socket
 * @args: RPC transport to connect
 *
 * Invoked by a work queue tasklet.
 */
static void xs_udp_connect_worker(void *args)
846
{
847 848 849
	struct rpc_xprt *xprt = (struct rpc_xprt *) args;
	struct socket *sock = xprt->sock;
	int err, status = -EIO;
850

851 852
	if (xprt->shutdown || xprt->addr.sin_port == 0)
		goto out;
853

854
	dprintk("RPC:      xs_udp_connect_worker for xprt %p\n", xprt);
855

856 857
	/* Start by resetting any existing state */
	xs_close(xprt);
858

859 860 861 862
	if ((err = sock_create_kern(PF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock)) < 0) {
		dprintk("RPC:      can't create UDP transport socket (%d).\n", -err);
		goto out;
	}
863

864 865 866 867
	if (xprt->resvport && xs_bindresvport(xprt, sock) < 0) {
		sock_release(sock);
		goto out;
	}
868

869 870
	if (!xprt->inet) {
		struct sock *sk = sock->sk;
871

872
		write_lock_bh(&sk->sk_callback_lock);
873

874 875 876 877
		sk->sk_user_data = xprt;
		xprt->old_data_ready = sk->sk_data_ready;
		xprt->old_state_change = sk->sk_state_change;
		xprt->old_write_space = sk->sk_write_space;
878
		sk->sk_data_ready = xs_udp_data_ready;
879
		sk->sk_write_space = xs_udp_write_space;
880
		sk->sk_no_check = UDP_CSUM_NORCV;
881

882 883
		xprt_set_connected(xprt);

884 885 886
		/* Reset to new socket */
		xprt->sock = sock;
		xprt->inet = sk;
887

888 889 890 891 892 893 894
		write_unlock_bh(&sk->sk_callback_lock);
	}
	xs_set_buffer_size(xprt);
	status = 0;
out:
	xprt_wake_pending_tasks(xprt, status);
	xprt_clear_connecting(xprt);
895 896
}

897
/**
898
 * xs_tcp_connect_worker - connect a TCP socket to a remote endpoint
899 900 901
 * @args: RPC transport to connect
 *
 * Invoked by a work queue tasklet.
902
 */
903
static void xs_tcp_connect_worker(void *args)
904 905 906
{
	struct rpc_xprt *xprt = (struct rpc_xprt *)args;
	struct socket *sock = xprt->sock;
907
	int err, status = -EIO;
908 909 910 911

	if (xprt->shutdown || xprt->addr.sin_port == 0)
		goto out;

912
	dprintk("RPC:      xs_tcp_connect_worker for xprt %p\n", xprt);
913

914
	/* Start by resetting any existing socket state */
915
	xs_close(xprt);
916 917 918

	if ((err = sock_create_kern(PF_INET, SOCK_STREAM, IPPROTO_TCP, &sock)) < 0) {
		dprintk("RPC:      can't create TCP transport socket (%d).\n", -err);
919 920 921
		goto out;
	}

922 923
	if (xprt->resvport && xs_bindresvport(xprt, sock) < 0) {
		sock_release(sock);
924
		goto out;
925
	}
926

927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950
	if (!xprt->inet) {
		struct sock *sk = sock->sk;

		write_lock_bh(&sk->sk_callback_lock);

		sk->sk_user_data = xprt;
		xprt->old_data_ready = sk->sk_data_ready;
		xprt->old_state_change = sk->sk_state_change;
		xprt->old_write_space = sk->sk_write_space;
		sk->sk_data_ready = xs_tcp_data_ready;
		sk->sk_state_change = xs_tcp_state_change;
		sk->sk_write_space = xs_tcp_write_space;
		tcp_sk(sk)->nonagle = 1;

		xprt_clear_connected(xprt);

		/* Reset to new socket */
		xprt->sock = sock;
		xprt->inet = sk;

		write_unlock_bh(&sk->sk_callback_lock);
	}

	/* Tell the socket layer to start connecting... */
951 952 953 954 955 956 957 958 959 960 961 962
	status = sock->ops->connect(sock, (struct sockaddr *) &xprt->addr,
			sizeof(xprt->addr), O_NONBLOCK);
	dprintk("RPC: %p  connect status %d connected %d sock state %d\n",
			xprt, -status, xprt_connected(xprt), sock->sk->sk_state);
	if (status < 0) {
		switch (status) {
			case -EINPROGRESS:
			case -EALREADY:
				goto out_clear;
		}
	}
out:
963
	xprt_wake_pending_tasks(xprt, status);
964
out_clear:
965
	xprt_clear_connecting(xprt);
966 967
}

968 969 970 971 972 973 974
/**
 * xs_connect - connect a socket to a remote endpoint
 * @task: address of RPC task that manages state of connect request
 *
 * TCP: If the remote end dropped the connection, delay reconnecting.
 */
static void xs_connect(struct rpc_task *task)
975 976 977
{
	struct rpc_xprt *xprt = task->tk_xprt;

978 979 980 981 982 983
	if (xprt_test_and_set_connecting(xprt))
		return;

	if (xprt->sock != NULL) {
		dprintk("RPC:      xs_connect delayed xprt %p\n", xprt);
		schedule_delayed_work(&xprt->connect_worker,
984
					RPC_REESTABLISH_TIMEOUT);
985 986 987 988 989 990 991
	} else {
		dprintk("RPC:      xs_connect scheduled xprt %p\n", xprt);
		schedule_work(&xprt->connect_worker);

		/* flush_scheduled_work can sleep... */
		if (!RPC_IS_ASYNC(task))
			flush_scheduled_work();
992 993 994
	}
}

995 996 997 998 999 1000
static struct rpc_xprt_ops xs_ops = {
	.set_buffer_size	= xs_set_buffer_size,
	.connect		= xs_connect,
	.send_request		= xs_send_request,
	.close			= xs_close,
	.destroy		= xs_destroy,
1001 1002 1003 1004 1005
};

extern unsigned int xprt_udp_slot_table_entries;
extern unsigned int xprt_tcp_slot_table_entries;

1006 1007 1008 1009 1010 1011
/**
 * xs_setup_udp - Set up transport to use a UDP socket
 * @xprt: transport to set up
 * @to:   timeout parameters
 *
 */
1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025
int xs_setup_udp(struct rpc_xprt *xprt, struct rpc_timeout *to)
{
	size_t slot_table_size;

	dprintk("RPC:      setting up udp-ipv4 transport...\n");

	xprt->max_reqs = xprt_udp_slot_table_entries;
	slot_table_size = xprt->max_reqs * sizeof(xprt->slot[0]);
	xprt->slot = kmalloc(slot_table_size, GFP_KERNEL);
	if (xprt->slot == NULL)
		return -ENOMEM;
	memset(xprt->slot, 0, slot_table_size);

	xprt->prot = IPPROTO_UDP;
1026
	xprt->port = XS_MAX_RESVPORT;
1027 1028 1029 1030 1031 1032 1033
	xprt->stream = 0;
	xprt->nocong = 0;
	xprt->cwnd = RPC_INITCWND;
	xprt->resvport = capable(CAP_NET_BIND_SERVICE) ? 1 : 0;
	/* XXX: header size can vary due to auth type, IPv6, etc. */
	xprt->max_payload = (1U << 16) - (MAX_HEADER << 3);

1034
	INIT_WORK(&xprt->connect_worker, xs_udp_connect_worker, xprt);
1035

1036
	xprt->ops = &xs_ops;
1037 1038 1039 1040

	if (to)
		xprt->timeout = *to;
	else
1041
		xprt_set_timeout(&xprt->timeout, 5, 5 * HZ);
1042 1043 1044 1045

	return 0;
}

1046 1047 1048 1049 1050 1051
/**
 * xs_setup_tcp - Set up transport to use a TCP socket
 * @xprt: transport to set up
 * @to: timeout parameters
 *
 */
1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065
int xs_setup_tcp(struct rpc_xprt *xprt, struct rpc_timeout *to)
{
	size_t slot_table_size;

	dprintk("RPC:      setting up tcp-ipv4 transport...\n");

	xprt->max_reqs = xprt_tcp_slot_table_entries;
	slot_table_size = xprt->max_reqs * sizeof(xprt->slot[0]);
	xprt->slot = kmalloc(slot_table_size, GFP_KERNEL);
	if (xprt->slot == NULL)
		return -ENOMEM;
	memset(xprt->slot, 0, slot_table_size);

	xprt->prot = IPPROTO_TCP;
1066
	xprt->port = XS_MAX_RESVPORT;
1067 1068 1069 1070 1071 1072
	xprt->stream = 1;
	xprt->nocong = 1;
	xprt->cwnd = RPC_MAXCWND(xprt);
	xprt->resvport = capable(CAP_NET_BIND_SERVICE) ? 1 : 0;
	xprt->max_payload = (1U << 31) - 1;

1073
	INIT_WORK(&xprt->connect_worker, xs_tcp_connect_worker, xprt);
1074

1075
	xprt->ops = &xs_ops;
1076 1077 1078 1079

	if (to)
		xprt->timeout = *to;
	else
1080
		xprt_set_timeout(&xprt->timeout, 2, 60 * HZ);
1081 1082 1083

	return 0;
}