channel_mgmt.c 23.7 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
/*
 * Copyright (c) 2009, Microsoft Corporation.
 *
 * This program is free software; you can redistribute it and/or modify it
 * under the terms and conditions of the GNU General Public License,
 * version 2, as published by the Free Software Foundation.
 *
 * This program is distributed in the hope it will be useful, but WITHOUT
 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
 * more details.
 *
 * You should have received a copy of the GNU General Public License along with
 * this program; if not, write to the Free Software Foundation, Inc., 59 Temple
 * Place - Suite 330, Boston, MA 02111-1307 USA.
 *
 * Authors:
 *   Haiyang Zhang <haiyangz@microsoft.com>
 *   Hank Janssen  <hjanssen@microsoft.com>
 */
21 22
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

23
#include <linux/kernel.h>
24 25
#include <linux/sched.h>
#include <linux/wait.h>
26
#include <linux/mm.h>
27
#include <linux/slab.h>
28
#include <linux/list.h>
29
#include <linux/module.h>
30
#include <linux/completion.h>
31
#include <linux/hyperv.h>
32

33
#include "hyperv_vmbus.h"
34

35
struct vmbus_channel_message_table_entry {
36
	enum vmbus_channel_message_type message_type;
37
	void (*message_handler)(struct vmbus_channel_message_header *msg);
38
};
39

40 41

/**
42
 * vmbus_prep_negotiate_resp() - Create default response for Hyper-V Negotiate message
43 44 45 46 47 48
 * @icmsghdrp: Pointer to msg header structure
 * @icmsg_negotiate: Pointer to negotiate message structure
 * @buf: Raw buffer channel data
 *
 * @icmsghdrp is of type &struct icmsg_hdr.
 * @negop is of type &struct icmsg_negotiate.
49 50
 * Set up and fill in default negotiate response message.
 *
51 52 53
 * The fw_version specifies the  framework version that
 * we can support and srv_version specifies the service
 * version we can support.
54 55 56
 *
 * Mainly used by Hyper-V drivers.
 */
57
bool vmbus_prep_negotiate_resp(struct icmsg_hdr *icmsghdrp,
58
				struct icmsg_negotiate *negop, u8 *buf,
59
				int fw_version, int srv_version)
60
{
61 62 63 64
	int icframe_major, icframe_minor;
	int icmsg_major, icmsg_minor;
	int fw_major, fw_minor;
	int srv_major, srv_minor;
65
	int i;
66
	bool found_match = false;
67

68
	icmsghdrp->icmsgsize = 0x10;
69 70 71 72 73
	fw_major = (fw_version >> 16);
	fw_minor = (fw_version & 0xFFFF);

	srv_major = (srv_version >> 16);
	srv_minor = (srv_version & 0xFFFF);
74

75 76 77
	negop = (struct icmsg_negotiate *)&buf[
		sizeof(struct vmbuspipe_hdr) +
		sizeof(struct icmsg_hdr)];
78

79 80 81 82 83
	icframe_major = negop->icframe_vercnt;
	icframe_minor = 0;

	icmsg_major = negop->icmsg_vercnt;
	icmsg_minor = 0;
84 85 86 87 88 89 90

	/*
	 * Select the framework version number we will
	 * support.
	 */

	for (i = 0; i < negop->icframe_vercnt; i++) {
91 92 93 94 95 96
		if ((negop->icversion_data[i].major == fw_major) &&
		   (negop->icversion_data[i].minor == fw_minor)) {
			icframe_major = negop->icversion_data[i].major;
			icframe_minor = negop->icversion_data[i].minor;
			found_match = true;
		}
97 98
	}

99 100 101 102 103
	if (!found_match)
		goto fw_error;

	found_match = false;

104 105
	for (i = negop->icframe_vercnt;
		 (i < negop->icframe_vercnt + negop->icmsg_vercnt); i++) {
106 107 108 109 110 111
		if ((negop->icversion_data[i].major == srv_major) &&
		   (negop->icversion_data[i].minor == srv_minor)) {
			icmsg_major = negop->icversion_data[i].major;
			icmsg_minor = negop->icversion_data[i].minor;
			found_match = true;
		}
112
	}
113

114
	/*
115
	 * Respond with the framework and service
116 117
	 * version numbers we can support.
	 */
118 119 120 121 122 123 124 125 126 127 128 129 130 131 132

fw_error:
	if (!found_match) {
		negop->icframe_vercnt = 0;
		negop->icmsg_vercnt = 0;
	} else {
		negop->icframe_vercnt = 1;
		negop->icmsg_vercnt = 1;
	}

	negop->icversion_data[0].major = icframe_major;
	negop->icversion_data[0].minor = icframe_minor;
	negop->icversion_data[1].major = icmsg_major;
	negop->icversion_data[1].minor = icmsg_minor;
	return found_match;
133
}
134

135
EXPORT_SYMBOL_GPL(vmbus_prep_negotiate_resp);
136

137
/*
138
 * alloc_channel - Allocate and initialize a vmbus channel object
139
 */
140
static struct vmbus_channel *alloc_channel(void)
141
{
142
	struct vmbus_channel *channel;
143

144
	channel = kzalloc(sizeof(*channel), GFP_ATOMIC);
145 146 147
	if (!channel)
		return NULL;

148
	spin_lock_init(&channel->inbound_lock);
149 150 151
	spin_lock_init(&channel->sc_lock);

	INIT_LIST_HEAD(&channel->sc_list);
152
	INIT_LIST_HEAD(&channel->percpu_list);
153

154 155
	channel->controlwq = create_workqueue("hv_vmbus_ctl");
	if (!channel->controlwq) {
156
		kfree(channel);
157 158 159 160 161 162
		return NULL;
	}

	return channel;
}

163
/*
164
 * release_hannel - Release the vmbus channel object itself
165
 */
166
static void release_channel(struct work_struct *work)
167
{
168 169 170
	struct vmbus_channel *channel = container_of(work,
						     struct vmbus_channel,
						     work);
171

172
	destroy_workqueue(channel->controlwq);
173

174
	kfree(channel);
175 176
}

177
/*
178
 * free_channel - Release the resources used by the vmbus channel object
179
 */
180
static void free_channel(struct vmbus_channel *channel)
181 182
{

183 184 185 186 187
	/*
	 * We have to release the channel's workqueue/thread in the vmbus's
	 * workqueue/thread context
	 * ie we can't destroy ourselves.
	 */
188
	INIT_WORK(&channel->work, release_channel);
189
	queue_work(vmbus_connection.work_queue, &channel->work);
190 191
}

192 193 194 195 196 197 198
static void percpu_channel_enq(void *arg)
{
	struct vmbus_channel *channel = arg;
	int cpu = smp_processor_id();

	list_add_tail(&channel->percpu_list, &hv_context.percpu_list[cpu]);
}
199

200 201 202 203 204 205
static void percpu_channel_deq(void *arg)
{
	struct vmbus_channel *channel = arg;

	list_del(&channel->percpu_list);
}
206

207 208 209 210 211 212 213 214 215
/*
 * vmbus_process_rescind_offer -
 * Rescind the offer by initiating a device removal
 */
static void vmbus_process_rescind_offer(struct work_struct *work)
{
	struct vmbus_channel *channel = container_of(work,
						     struct vmbus_channel,
						     work);
216
	unsigned long flags;
217
	struct vmbus_channel *primary_channel;
218
	struct vmbus_channel_relid_released msg;
219 220 221 222 223 224 225 226 227
	struct device *dev;

	if (channel->device_obj) {
		dev = get_device(&channel->device_obj->device);
		if (dev) {
			vmbus_device_unregister(channel->device_obj);
			put_device(dev);
		}
	}
228

229 230 231 232 233
	memset(&msg, 0, sizeof(struct vmbus_channel_relid_released));
	msg.child_relid = channel->offermsg.child_relid;
	msg.header.msgtype = CHANNELMSG_RELID_RELEASED;
	vmbus_post_msg(&msg, sizeof(struct vmbus_channel_relid_released));

234 235
	if (channel->target_cpu != get_cpu()) {
		put_cpu();
236 237
		smp_call_function_single(channel->target_cpu,
					 percpu_channel_deq, channel, true);
238
	} else {
239
		percpu_channel_deq(channel);
240 241
		put_cpu();
	}
242

243 244 245 246 247 248 249
	if (channel->primary_channel == NULL) {
		spin_lock_irqsave(&vmbus_connection.channel_lock, flags);
		list_del(&channel->listentry);
		spin_unlock_irqrestore(&vmbus_connection.channel_lock, flags);
	} else {
		primary_channel = channel->primary_channel;
		spin_lock_irqsave(&primary_channel->sc_lock, flags);
250
		list_del(&channel->sc_list);
251 252
		spin_unlock_irqrestore(&primary_channel->sc_lock, flags);
	}
253
	free_channel(channel);
254
}
255

256 257 258 259 260 261 262 263 264 265 266
void vmbus_free_channels(void)
{
	struct vmbus_channel *channel;

	list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
		vmbus_device_unregister(channel->device_obj);
		kfree(channel->device_obj);
		free_channel(channel);
	}
}

267
/*
268
 * vmbus_process_offer - Process the offer by creating a channel/device
269
 * associated with this offer
270
 */
271
static void vmbus_process_offer(struct work_struct *work)
272
{
273 274 275
	struct vmbus_channel *newchannel = container_of(work,
							struct vmbus_channel,
							work);
276
	struct vmbus_channel *channel;
277
	bool fnew = true;
278
	bool enq = false;
279
	int ret;
280
	unsigned long flags;
281

282 283 284
	/* The next possible work is rescind handling */
	INIT_WORK(&newchannel->work, vmbus_process_rescind_offer);

285
	/* Make sure this is a new offer */
286
	spin_lock_irqsave(&vmbus_connection.channel_lock, flags);
287

288
	list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
289 290 291 292
		if (!uuid_le_cmp(channel->offermsg.offer.if_type,
			newchannel->offermsg.offer.if_type) &&
			!uuid_le_cmp(channel->offermsg.offer.if_instance,
				newchannel->offermsg.offer.if_instance)) {
293
			fnew = false;
294 295 296 297
			break;
		}
	}

298
	if (fnew) {
299
		list_add_tail(&newchannel->listentry,
300
			      &vmbus_connection.chn_list);
301 302
		enq = true;
	}
303

304
	spin_unlock_irqrestore(&vmbus_connection.channel_lock, flags);
305

306
	if (enq) {
307 308
		if (newchannel->target_cpu != get_cpu()) {
			put_cpu();
309 310 311
			smp_call_function_single(newchannel->target_cpu,
						 percpu_channel_enq,
						 newchannel, true);
312
		} else {
313
			percpu_channel_enq(newchannel);
314 315
			put_cpu();
		}
316
	}
317
	if (!fnew) {
318 319 320 321 322 323 324 325 326 327 328
		/*
		 * Check to see if this is a sub-channel.
		 */
		if (newchannel->offermsg.offer.sub_channel_index != 0) {
			/*
			 * Process the sub-channel.
			 */
			newchannel->primary_channel = channel;
			spin_lock_irqsave(&channel->sc_lock, flags);
			list_add_tail(&newchannel->sc_list, &channel->sc_list);
			spin_unlock_irqrestore(&channel->sc_lock, flags);
329

330 331
			if (newchannel->target_cpu != get_cpu()) {
				put_cpu();
332 333 334
				smp_call_function_single(newchannel->target_cpu,
							 percpu_channel_enq,
							 newchannel, true);
335
			} else {
336
				percpu_channel_enq(newchannel);
337 338
				put_cpu();
			}
339

340 341 342 343
			newchannel->state = CHANNEL_OPEN_STATE;
			if (channel->sc_creation_callback != NULL)
				channel->sc_creation_callback(newchannel);

344
			goto out;
345 346
		}

347
		goto err_free_chan;
348 349
	}

350 351 352 353 354 355 356
	/*
	 * This state is used to indicate a successful open
	 * so that when we do close the channel normally, we
	 * can cleanup properly
	 */
	newchannel->state = CHANNEL_OPEN_STATE;

357 358 359
	/*
	 * Start the process of binding this offer to the driver
	 * We need to set the DeviceObject field before calling
360
	 * vmbus_child_dev_add()
361
	 */
362
	newchannel->device_obj = vmbus_device_create(
363 364
		&newchannel->offermsg.offer.if_type,
		&newchannel->offermsg.offer.if_instance,
365
		newchannel);
366 367
	if (!newchannel->device_obj)
		goto err_free_chan;
368

369 370 371 372 373
	/*
	 * Add the new device to the bus. This will kick off device-driver
	 * binding which eventually invokes the device driver's AddDevice()
	 * method.
	 */
374
	ret = vmbus_device_register(newchannel->device_obj);
375
	if (ret != 0) {
376
		pr_err("unable to add child device object (relid %d)\n",
377
			   newchannel->offermsg.child_relid);
378

379
		spin_lock_irqsave(&vmbus_connection.channel_lock, flags);
380
		list_del(&newchannel->listentry);
381
		spin_unlock_irqrestore(&vmbus_connection.channel_lock, flags);
K
K. Y. Srinivasan 已提交
382
		kfree(newchannel->device_obj);
383
		goto err_free_chan;
384
	}
385 386 387 388
out:
	return;
err_free_chan:
	free_channel(newchannel);
389 390
}

391 392 393 394 395 396 397 398
enum {
	IDE = 0,
	SCSI,
	NIC,
	MAX_PERF_CHN,
};

/*
399
 * This is an array of device_ids (device types) that are performance critical.
400 401 402
 * We attempt to distribute the interrupt load for these devices across
 * all available CPUs.
 */
403
static const struct hv_vmbus_device_id hp_devs[] = {
404
	/* IDE */
405
	{ HV_IDE_GUID, },
406
	/* Storage - SCSI */
407
	{ HV_SCSI_GUID, },
408
	/* Network */
409
	{ HV_NIC_GUID, },
410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426
};


/*
 * We use this state to statically distribute the channel interrupt load.
 */
static u32  next_vp;

/*
 * Starting with Win8, we can statically distribute the incoming
 * channel interrupt load by binding a channel to VCPU. We
 * implement here a simple round robin scheme for distributing
 * the interrupt load.
 * We will bind channels that are not performance critical to cpu 0 and
 * performance critical channels (IDE, SCSI and Network) will be uniformly
 * distributed across all available CPUs.
 */
427
static void init_vp_index(struct vmbus_channel *channel, const uuid_le *type_guid)
428 429 430 431 432 433 434
{
	u32 cur_cpu;
	int i;
	bool perf_chn = false;
	u32 max_cpus = num_online_cpus();

	for (i = IDE; i < MAX_PERF_CHN; i++) {
435
		if (!memcmp(type_guid->b, hp_devs[i].guid,
436 437 438 439 440 441 442 443 444 445 446 447 448
				 sizeof(uuid_le))) {
			perf_chn = true;
			break;
		}
	}
	if ((vmbus_proto_version == VERSION_WS2008) ||
	    (vmbus_proto_version == VERSION_WIN7) || (!perf_chn)) {
		/*
		 * Prior to win8, all channel interrupts are
		 * delivered on cpu 0.
		 * Also if the channel is not a performance critical
		 * channel, bind it to cpu 0.
		 */
449 450 451
		channel->target_cpu = 0;
		channel->target_vp = 0;
		return;
452 453
	}
	cur_cpu = (++next_vp % max_cpus);
454 455
	channel->target_cpu = cur_cpu;
	channel->target_vp = hv_context.vp_index[cur_cpu];
456 457
}

458
/*
459
 * vmbus_onoffer - Handler for channel offers from vmbus in parent partition.
460 461
 *
 */
462
static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
463
{
464
	struct vmbus_channel_offer_channel *offer;
465
	struct vmbus_channel *newchannel;
466

467
	offer = (struct vmbus_channel_offer_channel *)hdr;
468

469
	/* Allocate the channel object and save this offer. */
470
	newchannel = alloc_channel();
471
	if (!newchannel) {
472
		pr_err("Unable to allocate channel object\n");
473 474 475
		return;
	}

476 477 478 479 480 481 482
	/*
	 * By default we setup state to enable batched
	 * reading. A specific service can choose to
	 * disable this prior to opening the channel.
	 */
	newchannel->batched_reading = true;

483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502
	/*
	 * Setup state for signalling the host.
	 */
	newchannel->sig_event = (struct hv_input_signal_event *)
				(ALIGN((unsigned long)
				&newchannel->sig_buf,
				HV_HYPERCALL_PARAM_ALIGN));

	newchannel->sig_event->connectionid.asu32 = 0;
	newchannel->sig_event->connectionid.u.id = VMBUS_EVENT_CONNECTION_ID;
	newchannel->sig_event->flag_number = 0;
	newchannel->sig_event->rsvdz = 0;

	if (vmbus_proto_version != VERSION_WS2008) {
		newchannel->is_dedicated_interrupt =
				(offer->is_dedicated_interrupt != 0);
		newchannel->sig_event->connectionid.u.id =
				offer->connection_id;
	}

503
	init_vp_index(newchannel, &offer->offer.if_type);
504

505
	memcpy(&newchannel->offermsg, offer,
506
	       sizeof(struct vmbus_channel_offer_channel));
507 508
	newchannel->monitor_grp = (u8)offer->monitorid / 32;
	newchannel->monitor_bit = (u8)offer->monitorid % 32;
509

510 511
	INIT_WORK(&newchannel->work, vmbus_process_offer);
	queue_work(newchannel->controlwq, &newchannel->work);
512 513
}

514
/*
515
 * vmbus_onoffer_rescind - Rescind offer handler.
516 517 518
 *
 * We queue a work item to process this offer synchronously
 */
519
static void vmbus_onoffer_rescind(struct vmbus_channel_message_header *hdr)
520
{
521
	struct vmbus_channel_rescind_offer *rescind;
522
	struct vmbus_channel *channel;
523

524
	rescind = (struct vmbus_channel_rescind_offer *)hdr;
525
	channel = relid2channel(rescind->child_relid);
526 527 528

	if (channel == NULL)
		/* Just return here, no channel found */
529 530
		return;

531 532
	channel->rescind = true;

533 534 535
	/* work is initialized for vmbus_process_rescind_offer() from
	 * vmbus_process_offer() where the channel got created */
	queue_work(channel->controlwq, &channel->work);
536 537
}

538
/*
539 540
 * vmbus_onoffers_delivered -
 * This is invoked when all offers have been delivered.
541 542 543
 *
 * Nothing to do here.
 */
544
static void vmbus_onoffers_delivered(
545
			struct vmbus_channel_message_header *hdr)
546 547 548
{
}

549
/*
550
 * vmbus_onopen_result - Open result handler.
551 552 553 554 555
 *
 * This is invoked when we received a response to our channel open request.
 * Find the matching request, copy the response and signal the requesting
 * thread.
 */
556
static void vmbus_onopen_result(struct vmbus_channel_message_header *hdr)
557
{
558
	struct vmbus_channel_open_result *result;
559 560 561
	struct vmbus_channel_msginfo *msginfo;
	struct vmbus_channel_message_header *requestheader;
	struct vmbus_channel_open_channel *openmsg;
562
	unsigned long flags;
563

564
	result = (struct vmbus_channel_open_result *)hdr;
565

566 567 568
	/*
	 * Find the open msg, copy the result and signal/unblock the wait event
	 */
569
	spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags);
570

571 572
	list_for_each_entry(msginfo, &vmbus_connection.chn_msg_list,
				msglistentry) {
573
		requestheader =
574
			(struct vmbus_channel_message_header *)msginfo->msg;
575

576
		if (requestheader->msgtype == CHANNELMSG_OPENCHANNEL) {
577
			openmsg =
578 579 580 581
			(struct vmbus_channel_open_channel *)msginfo->msg;
			if (openmsg->child_relid == result->child_relid &&
			    openmsg->openid == result->openid) {
				memcpy(&msginfo->response.open_result,
582
				       result,
583 584 585
				       sizeof(
					struct vmbus_channel_open_result));
				complete(&msginfo->waitevent);
586 587 588 589
				break;
			}
		}
	}
590
	spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags);
591 592
}

593
/*
594
 * vmbus_ongpadl_created - GPADL created handler.
595 596 597 598 599
 *
 * This is invoked when we received a response to our gpadl create request.
 * Find the matching request, copy the response and signal the requesting
 * thread.
 */
600
static void vmbus_ongpadl_created(struct vmbus_channel_message_header *hdr)
601
{
602 603 604 605
	struct vmbus_channel_gpadl_created *gpadlcreated;
	struct vmbus_channel_msginfo *msginfo;
	struct vmbus_channel_message_header *requestheader;
	struct vmbus_channel_gpadl_header *gpadlheader;
606
	unsigned long flags;
607

608
	gpadlcreated = (struct vmbus_channel_gpadl_created *)hdr;
609

610 611 612 613
	/*
	 * Find the establish msg, copy the result and signal/unblock the wait
	 * event
	 */
614
	spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags);
615

616 617
	list_for_each_entry(msginfo, &vmbus_connection.chn_msg_list,
				msglistentry) {
618
		requestheader =
619
			(struct vmbus_channel_message_header *)msginfo->msg;
620

621
		if (requestheader->msgtype == CHANNELMSG_GPADL_HEADER) {
622 623 624
			gpadlheader =
			(struct vmbus_channel_gpadl_header *)requestheader;

625 626 627 628
			if ((gpadlcreated->child_relid ==
			     gpadlheader->child_relid) &&
			    (gpadlcreated->gpadl == gpadlheader->gpadl)) {
				memcpy(&msginfo->response.gpadl_created,
629
				       gpadlcreated,
630 631 632
				       sizeof(
					struct vmbus_channel_gpadl_created));
				complete(&msginfo->waitevent);
633 634 635 636
				break;
			}
		}
	}
637
	spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags);
638 639
}

640
/*
641
 * vmbus_ongpadl_torndown - GPADL torndown handler.
642 643 644 645 646
 *
 * This is invoked when we received a response to our gpadl teardown request.
 * Find the matching request, copy the response and signal the requesting
 * thread.
 */
647
static void vmbus_ongpadl_torndown(
648
			struct vmbus_channel_message_header *hdr)
649
{
650 651 652 653
	struct vmbus_channel_gpadl_torndown *gpadl_torndown;
	struct vmbus_channel_msginfo *msginfo;
	struct vmbus_channel_message_header *requestheader;
	struct vmbus_channel_gpadl_teardown *gpadl_teardown;
654
	unsigned long flags;
655

656
	gpadl_torndown = (struct vmbus_channel_gpadl_torndown *)hdr;
657 658 659 660

	/*
	 * Find the open msg, copy the result and signal/unblock the wait event
	 */
661
	spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags);
662

663 664
	list_for_each_entry(msginfo, &vmbus_connection.chn_msg_list,
				msglistentry) {
665
		requestheader =
666
			(struct vmbus_channel_message_header *)msginfo->msg;
667

668
		if (requestheader->msgtype == CHANNELMSG_GPADL_TEARDOWN) {
669 670
			gpadl_teardown =
			(struct vmbus_channel_gpadl_teardown *)requestheader;
671

672 673
			if (gpadl_torndown->gpadl == gpadl_teardown->gpadl) {
				memcpy(&msginfo->response.gpadl_torndown,
674
				       gpadl_torndown,
675 676 677
				       sizeof(
					struct vmbus_channel_gpadl_torndown));
				complete(&msginfo->waitevent);
678 679 680 681
				break;
			}
		}
	}
682
	spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags);
683 684
}

685
/*
686
 * vmbus_onversion_response - Version response handler
687 688 689 690 691
 *
 * This is invoked when we received a response to our initiate contact request.
 * Find the matching request, copy the response and signal the requesting
 * thread.
 */
692
static void vmbus_onversion_response(
693
		struct vmbus_channel_message_header *hdr)
694
{
695 696 697
	struct vmbus_channel_msginfo *msginfo;
	struct vmbus_channel_message_header *requestheader;
	struct vmbus_channel_version_response *version_response;
698
	unsigned long flags;
699

700
	version_response = (struct vmbus_channel_version_response *)hdr;
701
	spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags);
702

703 704
	list_for_each_entry(msginfo, &vmbus_connection.chn_msg_list,
				msglistentry) {
705
		requestheader =
706
			(struct vmbus_channel_message_header *)msginfo->msg;
707

708 709 710
		if (requestheader->msgtype ==
		    CHANNELMSG_INITIATE_CONTACT) {
			memcpy(&msginfo->response.version_response,
711
			      version_response,
712
			      sizeof(struct vmbus_channel_version_response));
713
			complete(&msginfo->waitevent);
714 715
		}
	}
716
	spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags);
717 718
}

719 720
/* Channel message dispatch table */
static struct vmbus_channel_message_table_entry
721
	channel_message_table[CHANNELMSG_COUNT] = {
722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738
	{CHANNELMSG_INVALID,			NULL},
	{CHANNELMSG_OFFERCHANNEL,		vmbus_onoffer},
	{CHANNELMSG_RESCIND_CHANNELOFFER,	vmbus_onoffer_rescind},
	{CHANNELMSG_REQUESTOFFERS,		NULL},
	{CHANNELMSG_ALLOFFERS_DELIVERED,	vmbus_onoffers_delivered},
	{CHANNELMSG_OPENCHANNEL,		NULL},
	{CHANNELMSG_OPENCHANNEL_RESULT,	vmbus_onopen_result},
	{CHANNELMSG_CLOSECHANNEL,		NULL},
	{CHANNELMSG_GPADL_HEADER,		NULL},
	{CHANNELMSG_GPADL_BODY,		NULL},
	{CHANNELMSG_GPADL_CREATED,		vmbus_ongpadl_created},
	{CHANNELMSG_GPADL_TEARDOWN,		NULL},
	{CHANNELMSG_GPADL_TORNDOWN,		vmbus_ongpadl_torndown},
	{CHANNELMSG_RELID_RELEASED,		NULL},
	{CHANNELMSG_INITIATE_CONTACT,		NULL},
	{CHANNELMSG_VERSION_RESPONSE,		vmbus_onversion_response},
	{CHANNELMSG_UNLOAD,			NULL},
739 740
};

741
/*
742
 * vmbus_onmessage - Handler for channel protocol messages.
743 744 745
 *
 * This is invoked in the vmbus worker thread context.
 */
746
void vmbus_onmessage(void *context)
747
{
748
	struct hv_message *msg = context;
749
	struct vmbus_channel_message_header *hdr;
750 751
	int size;

752 753
	hdr = (struct vmbus_channel_message_header *)msg->u.payload;
	size = msg->header.payload_size;
754

755
	if (hdr->msgtype >= CHANNELMSG_COUNT) {
756
		pr_err("Received invalid channel message type %d size %d\n",
757
			   hdr->msgtype, size);
758
		print_hex_dump_bytes("", DUMP_PREFIX_NONE,
759
				     (unsigned char *)msg->u.payload, size);
760 761 762
		return;
	}

763 764
	if (channel_message_table[hdr->msgtype].message_handler)
		channel_message_table[hdr->msgtype].message_handler(hdr);
765
	else
766
		pr_err("Unhandled channel message type %d\n", hdr->msgtype);
767 768
}

769
/*
770
 * vmbus_request_offers - Send a request to get all our pending offers.
771
 */
772
int vmbus_request_offers(void)
773
{
774
	struct vmbus_channel_message_header *msg;
775
	struct vmbus_channel_msginfo *msginfo;
776
	int ret, t;
777

778
	msginfo = kmalloc(sizeof(*msginfo) +
779 780
			  sizeof(struct vmbus_channel_message_header),
			  GFP_KERNEL);
781
	if (!msginfo)
782
		return -ENOMEM;
783

784
	init_completion(&msginfo->waitevent);
785

786
	msg = (struct vmbus_channel_message_header *)msginfo->msg;
787

788
	msg->msgtype = CHANNELMSG_REQUESTOFFERS;
789 790


791
	ret = vmbus_post_msg(msg,
792 793
			       sizeof(struct vmbus_channel_message_header));
	if (ret != 0) {
794
		pr_err("Unable to request offers - %d\n", ret);
795

796 797
		goto cleanup;
	}
798

799
	t = wait_for_completion_timeout(&msginfo->waitevent, 5*HZ);
800
	if (t == 0) {
801 802
		ret = -ETIMEDOUT;
		goto cleanup;
803 804 805 806
	}



807
cleanup:
808
	kfree(msginfo);
809 810 811 812

	return ret;
}

813 814 815 816 817 818 819 820 821
/*
 * Retrieve the (sub) channel on which to send an outgoing request.
 * When a primary channel has multiple sub-channels, we choose a
 * channel whose VCPU binding is closest to the VCPU on which
 * this call is being made.
 */
struct vmbus_channel *vmbus_get_outgoing_channel(struct vmbus_channel *primary)
{
	struct list_head *cur, *tmp;
822
	int cur_cpu;
823 824 825 826 827 828 829
	struct vmbus_channel *cur_channel;
	struct vmbus_channel *outgoing_channel = primary;
	int cpu_distance, new_cpu_distance;

	if (list_empty(&primary->sc_list))
		return outgoing_channel;

830 831
	cur_cpu = hv_context.vp_index[get_cpu()];
	put_cpu();
832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897
	list_for_each_safe(cur, tmp, &primary->sc_list) {
		cur_channel = list_entry(cur, struct vmbus_channel, sc_list);
		if (cur_channel->state != CHANNEL_OPENED_STATE)
			continue;

		if (cur_channel->target_vp == cur_cpu)
			return cur_channel;

		cpu_distance = ((outgoing_channel->target_vp > cur_cpu) ?
				(outgoing_channel->target_vp - cur_cpu) :
				(cur_cpu - outgoing_channel->target_vp));

		new_cpu_distance = ((cur_channel->target_vp > cur_cpu) ?
				(cur_channel->target_vp - cur_cpu) :
				(cur_cpu - cur_channel->target_vp));

		if (cpu_distance < new_cpu_distance)
			continue;

		outgoing_channel = cur_channel;
	}

	return outgoing_channel;
}
EXPORT_SYMBOL_GPL(vmbus_get_outgoing_channel);

static void invoke_sc_cb(struct vmbus_channel *primary_channel)
{
	struct list_head *cur, *tmp;
	struct vmbus_channel *cur_channel;

	if (primary_channel->sc_creation_callback == NULL)
		return;

	list_for_each_safe(cur, tmp, &primary_channel->sc_list) {
		cur_channel = list_entry(cur, struct vmbus_channel, sc_list);

		primary_channel->sc_creation_callback(cur_channel);
	}
}

void vmbus_set_sc_create_callback(struct vmbus_channel *primary_channel,
				void (*sc_cr_cb)(struct vmbus_channel *new_sc))
{
	primary_channel->sc_creation_callback = sc_cr_cb;
}
EXPORT_SYMBOL_GPL(vmbus_set_sc_create_callback);

bool vmbus_are_subchannels_present(struct vmbus_channel *primary)
{
	bool ret;

	ret = !list_empty(&primary->sc_list);

	if (ret) {
		/*
		 * Invoke the callback on sub-channel creation.
		 * This will present a uniform interface to the
		 * clients.
		 */
		invoke_sc_cb(primary);
	}

	return ret;
}
EXPORT_SYMBOL_GPL(vmbus_are_subchannels_present);