1 /*- 2 * Copyright (c) 2009-2012,2016 Microsoft Corp. 3 * Copyright (c) 2012 NetApp Inc. 4 * Copyright (c) 2012 Citrix Inc. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice unmodified, this list of conditions, and the following 12 * disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 27 */ 28 29 #include <sys/cdefs.h> 30 __FBSDID("$FreeBSD$"); 31 32 #include <sys/param.h> 33 #include <sys/bus.h> 34 #include <sys/kernel.h> 35 #include <sys/lock.h> 36 #include <sys/malloc.h> 37 #include <sys/mutex.h> 38 #include <sys/smp.h> 39 #include <sys/sysctl.h> 40 #include <sys/systm.h> 41 42 #include <machine/atomic.h> 43 #include <machine/stdarg.h> 44 45 #include <dev/hyperv/include/hyperv_busdma.h> 46 #include <dev/hyperv/include/vmbus_xact.h> 47 #include <dev/hyperv/vmbus/hyperv_var.h> 48 #include <dev/hyperv/vmbus/vmbus_reg.h> 49 #include <dev/hyperv/vmbus/vmbus_var.h> 50 #include <dev/hyperv/vmbus/vmbus_brvar.h> 51 #include <dev/hyperv/vmbus/vmbus_chanvar.h> 52 53 static void vmbus_chan_update_evtflagcnt( 54 struct vmbus_softc *, 55 const struct vmbus_channel *); 56 static int vmbus_chan_close_internal( 57 struct vmbus_channel *); 58 static int vmbus_chan_sysctl_mnf(SYSCTL_HANDLER_ARGS); 59 static void vmbus_chan_sysctl_create( 60 struct vmbus_channel *); 61 static struct vmbus_channel *vmbus_chan_alloc(struct vmbus_softc *); 62 static void vmbus_chan_free(struct vmbus_channel *); 63 static int vmbus_chan_add(struct vmbus_channel *); 64 static void vmbus_chan_cpu_default(struct vmbus_channel *); 65 static int vmbus_chan_release(struct vmbus_channel *); 66 static void vmbus_chan_set_chmap(struct vmbus_channel *); 67 static void vmbus_chan_clear_chmap(struct vmbus_channel *); 68 static void vmbus_chan_detach(struct vmbus_channel *); 69 static bool vmbus_chan_wait_revoke( 70 const struct vmbus_channel *, bool); 71 72 static void vmbus_chan_ins_prilist(struct vmbus_softc *, 73 struct vmbus_channel *); 74 static void vmbus_chan_rem_prilist(struct vmbus_softc *, 75 struct vmbus_channel *); 76 static void vmbus_chan_ins_list(struct vmbus_softc *, 77 struct vmbus_channel *); 78 static void vmbus_chan_rem_list(struct vmbus_softc *, 79 struct vmbus_channel *); 80 static void vmbus_chan_ins_sublist(struct vmbus_channel *, 81 struct vmbus_channel *); 82 static void vmbus_chan_rem_sublist(struct vmbus_channel *, 83 struct vmbus_channel *); 84 85 static void vmbus_chan_task(void *, int); 86 static void vmbus_chan_task_nobatch(void *, int); 87 static void vmbus_chan_clrchmap_task(void *, int); 88 static void vmbus_prichan_attach_task(void *, int); 89 static void vmbus_subchan_attach_task(void *, int); 90 static void vmbus_prichan_detach_task(void *, int); 91 static void vmbus_subchan_detach_task(void *, int); 92 93 static void vmbus_chan_msgproc_choffer(struct vmbus_softc *, 94 const struct vmbus_message *); 95 static void vmbus_chan_msgproc_chrescind( 96 struct vmbus_softc *, 97 const struct vmbus_message *); 98 99 static int vmbus_chan_printf(const struct vmbus_channel *, 100 const char *, ...) __printflike(2, 3); 101 102 /* 103 * Vmbus channel message processing. 104 */ 105 static const vmbus_chanmsg_proc_t 106 vmbus_chan_msgprocs[VMBUS_CHANMSG_TYPE_MAX] = { 107 VMBUS_CHANMSG_PROC(CHOFFER, vmbus_chan_msgproc_choffer), 108 VMBUS_CHANMSG_PROC(CHRESCIND, vmbus_chan_msgproc_chrescind), 109 110 VMBUS_CHANMSG_PROC_WAKEUP(CHOPEN_RESP), 111 VMBUS_CHANMSG_PROC_WAKEUP(GPADL_CONNRESP), 112 VMBUS_CHANMSG_PROC_WAKEUP(GPADL_DISCONNRESP) 113 }; 114 115 /* 116 * Notify host that there are data pending on our TX bufring. 117 */ 118 static __inline void 119 vmbus_chan_signal_tx(const struct vmbus_channel *chan) 120 { 121 atomic_set_long(chan->ch_evtflag, chan->ch_evtflag_mask); 122 if (chan->ch_txflags & VMBUS_CHAN_TXF_HASMNF) 123 atomic_set_int(chan->ch_montrig, chan->ch_montrig_mask); 124 else 125 hypercall_signal_event(chan->ch_monprm_dma.hv_paddr); 126 } 127 128 static void 129 vmbus_chan_ins_prilist(struct vmbus_softc *sc, struct vmbus_channel *chan) 130 { 131 132 mtx_assert(&sc->vmbus_prichan_lock, MA_OWNED); 133 if (atomic_testandset_int(&chan->ch_stflags, 134 VMBUS_CHAN_ST_ONPRIL_SHIFT)) 135 panic("channel is already on the prilist"); 136 TAILQ_INSERT_TAIL(&sc->vmbus_prichans, chan, ch_prilink); 137 } 138 139 static void 140 vmbus_chan_rem_prilist(struct vmbus_softc *sc, struct vmbus_channel *chan) 141 { 142 143 mtx_assert(&sc->vmbus_prichan_lock, MA_OWNED); 144 if (atomic_testandclear_int(&chan->ch_stflags, 145 VMBUS_CHAN_ST_ONPRIL_SHIFT) == 0) 146 panic("channel is not on the prilist"); 147 TAILQ_REMOVE(&sc->vmbus_prichans, chan, ch_prilink); 148 } 149 150 static void 151 vmbus_chan_ins_sublist(struct vmbus_channel *prichan, 152 struct vmbus_channel *chan) 153 { 154 155 mtx_assert(&prichan->ch_subchan_lock, MA_OWNED); 156 157 if (atomic_testandset_int(&chan->ch_stflags, 158 VMBUS_CHAN_ST_ONSUBL_SHIFT)) 159 panic("channel is already on the sublist"); 160 TAILQ_INSERT_TAIL(&prichan->ch_subchans, chan, ch_sublink); 161 162 /* Bump sub-channel count. */ 163 prichan->ch_subchan_cnt++; 164 } 165 166 static void 167 vmbus_chan_rem_sublist(struct vmbus_channel *prichan, 168 struct vmbus_channel *chan) 169 { 170 171 mtx_assert(&prichan->ch_subchan_lock, MA_OWNED); 172 173 KASSERT(prichan->ch_subchan_cnt > 0, 174 ("invalid subchan_cnt %d", prichan->ch_subchan_cnt)); 175 prichan->ch_subchan_cnt--; 176 177 if (atomic_testandclear_int(&chan->ch_stflags, 178 VMBUS_CHAN_ST_ONSUBL_SHIFT) == 0) 179 panic("channel is not on the sublist"); 180 TAILQ_REMOVE(&prichan->ch_subchans, chan, ch_sublink); 181 } 182 183 static void 184 vmbus_chan_ins_list(struct vmbus_softc *sc, struct vmbus_channel *chan) 185 { 186 187 mtx_assert(&sc->vmbus_chan_lock, MA_OWNED); 188 if (atomic_testandset_int(&chan->ch_stflags, 189 VMBUS_CHAN_ST_ONLIST_SHIFT)) 190 panic("channel is already on the list"); 191 TAILQ_INSERT_TAIL(&sc->vmbus_chans, chan, ch_link); 192 } 193 194 static void 195 vmbus_chan_rem_list(struct vmbus_softc *sc, struct vmbus_channel *chan) 196 { 197 198 mtx_assert(&sc->vmbus_chan_lock, MA_OWNED); 199 if (atomic_testandclear_int(&chan->ch_stflags, 200 VMBUS_CHAN_ST_ONLIST_SHIFT) == 0) 201 panic("channel is not on the list"); 202 TAILQ_REMOVE(&sc->vmbus_chans, chan, ch_link); 203 } 204 205 static int 206 vmbus_chan_sysctl_mnf(SYSCTL_HANDLER_ARGS) 207 { 208 struct vmbus_channel *chan = arg1; 209 int mnf = 0; 210 211 if (chan->ch_txflags & VMBUS_CHAN_TXF_HASMNF) 212 mnf = 1; 213 return sysctl_handle_int(oidp, &mnf, 0, req); 214 } 215 216 static void 217 vmbus_chan_sysctl_create(struct vmbus_channel *chan) 218 { 219 struct sysctl_oid *ch_tree, *chid_tree, *br_tree; 220 struct sysctl_ctx_list *ctx; 221 uint32_t ch_id; 222 char name[16]; 223 224 /* 225 * Add sysctl nodes related to this channel to this 226 * channel's sysctl ctx, so that they can be destroyed 227 * independently upon close of this channel, which can 228 * happen even if the device is not detached. 229 */ 230 ctx = &chan->ch_sysctl_ctx; 231 sysctl_ctx_init(ctx); 232 233 /* 234 * Create dev.NAME.UNIT.channel tree. 235 */ 236 ch_tree = SYSCTL_ADD_NODE(ctx, 237 SYSCTL_CHILDREN(device_get_sysctl_tree(chan->ch_dev)), 238 OID_AUTO, "channel", CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 239 if (ch_tree == NULL) 240 return; 241 242 /* 243 * Create dev.NAME.UNIT.channel.CHANID tree. 244 */ 245 if (VMBUS_CHAN_ISPRIMARY(chan)) 246 ch_id = chan->ch_id; 247 else 248 ch_id = chan->ch_prichan->ch_id; 249 snprintf(name, sizeof(name), "%d", ch_id); 250 chid_tree = SYSCTL_ADD_NODE(ctx, SYSCTL_CHILDREN(ch_tree), 251 OID_AUTO, name, CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 252 if (chid_tree == NULL) 253 return; 254 255 if (!VMBUS_CHAN_ISPRIMARY(chan)) { 256 /* 257 * Create dev.NAME.UNIT.channel.CHANID.sub tree. 258 */ 259 ch_tree = SYSCTL_ADD_NODE(ctx, SYSCTL_CHILDREN(chid_tree), 260 OID_AUTO, "sub", CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 261 if (ch_tree == NULL) 262 return; 263 264 /* 265 * Create dev.NAME.UNIT.channel.CHANID.sub.SUBIDX tree. 266 * 267 * NOTE: 268 * chid_tree is changed to this new sysctl tree. 269 */ 270 snprintf(name, sizeof(name), "%d", chan->ch_subidx); 271 chid_tree = SYSCTL_ADD_NODE(ctx, SYSCTL_CHILDREN(ch_tree), 272 OID_AUTO, name, CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 273 if (chid_tree == NULL) 274 return; 275 276 SYSCTL_ADD_UINT(ctx, SYSCTL_CHILDREN(chid_tree), OID_AUTO, 277 "chanid", CTLFLAG_RD, &chan->ch_id, 0, "channel id"); 278 } 279 280 SYSCTL_ADD_UINT(ctx, SYSCTL_CHILDREN(chid_tree), OID_AUTO, 281 "cpu", CTLFLAG_RD, &chan->ch_cpuid, 0, "owner CPU id"); 282 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(chid_tree), OID_AUTO, 283 "mnf", CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MPSAFE, 284 chan, 0, vmbus_chan_sysctl_mnf, "I", 285 "has monitor notification facilities"); 286 287 br_tree = SYSCTL_ADD_NODE(ctx, SYSCTL_CHILDREN(chid_tree), OID_AUTO, 288 "br", CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 289 if (br_tree != NULL) { 290 /* 291 * Create sysctl tree for RX bufring. 292 */ 293 vmbus_br_sysctl_create(ctx, br_tree, &chan->ch_rxbr.rxbr, "rx"); 294 /* 295 * Create sysctl tree for TX bufring. 296 */ 297 vmbus_br_sysctl_create(ctx, br_tree, &chan->ch_txbr.txbr, "tx"); 298 } 299 } 300 301 int 302 vmbus_chan_open(struct vmbus_channel *chan, int txbr_size, int rxbr_size, 303 const void *udata, int udlen, vmbus_chan_callback_t cb, void *cbarg) 304 { 305 struct vmbus_chan_br cbr; 306 int error; 307 308 /* 309 * Allocate the TX+RX bufrings. 310 */ 311 KASSERT(chan->ch_bufring == NULL, ("bufrings are allocated")); 312 chan->ch_bufring = hyperv_dmamem_alloc(bus_get_dma_tag(chan->ch_dev), 313 PAGE_SIZE, 0, txbr_size + rxbr_size, &chan->ch_bufring_dma, 314 BUS_DMA_WAITOK); 315 if (chan->ch_bufring == NULL) { 316 vmbus_chan_printf(chan, "bufring allocation failed\n"); 317 return (ENOMEM); 318 } 319 320 cbr.cbr = chan->ch_bufring; 321 cbr.cbr_paddr = chan->ch_bufring_dma.hv_paddr; 322 cbr.cbr_txsz = txbr_size; 323 cbr.cbr_rxsz = rxbr_size; 324 325 error = vmbus_chan_open_br(chan, &cbr, udata, udlen, cb, cbarg); 326 if (error) { 327 if (error == EISCONN) { 328 /* 329 * XXX 330 * The bufring GPADL is still connected; abandon 331 * this bufring, instead of having mysterious 332 * crash or trashed data later on. 333 */ 334 vmbus_chan_printf(chan, "chan%u bufring GPADL " 335 "is still connected upon channel open error; " 336 "leak %d bytes memory\n", chan->ch_id, 337 txbr_size + rxbr_size); 338 } else { 339 hyperv_dmamem_free(&chan->ch_bufring_dma, 340 chan->ch_bufring); 341 } 342 chan->ch_bufring = NULL; 343 } 344 return (error); 345 } 346 347 int 348 vmbus_chan_open_br(struct vmbus_channel *chan, const struct vmbus_chan_br *cbr, 349 const void *udata, int udlen, vmbus_chan_callback_t cb, void *cbarg) 350 { 351 struct vmbus_softc *sc = chan->ch_vmbus; 352 const struct vmbus_message *msg; 353 struct vmbus_chanmsg_chopen *req; 354 struct vmbus_msghc *mh; 355 uint32_t status; 356 int error, txbr_size, rxbr_size; 357 task_fn_t *task_fn; 358 uint8_t *br; 359 360 if (udlen > VMBUS_CHANMSG_CHOPEN_UDATA_SIZE) { 361 vmbus_chan_printf(chan, 362 "invalid udata len %d for chan%u\n", udlen, chan->ch_id); 363 return (EINVAL); 364 } 365 366 br = cbr->cbr; 367 txbr_size = cbr->cbr_txsz; 368 rxbr_size = cbr->cbr_rxsz; 369 KASSERT((txbr_size & PAGE_MASK) == 0, 370 ("send bufring size is not multiple page")); 371 KASSERT((rxbr_size & PAGE_MASK) == 0, 372 ("recv bufring size is not multiple page")); 373 KASSERT((cbr->cbr_paddr & PAGE_MASK) == 0, 374 ("bufring is not page aligned")); 375 376 /* 377 * Zero out the TX/RX bufrings, in case that they were used before. 378 */ 379 memset(br, 0, txbr_size + rxbr_size); 380 381 if (atomic_testandset_int(&chan->ch_stflags, 382 VMBUS_CHAN_ST_OPENED_SHIFT)) 383 panic("double-open chan%u", chan->ch_id); 384 385 chan->ch_cb = cb; 386 chan->ch_cbarg = cbarg; 387 388 vmbus_chan_update_evtflagcnt(sc, chan); 389 390 chan->ch_tq = VMBUS_PCPU_GET(chan->ch_vmbus, event_tq, chan->ch_cpuid); 391 if (chan->ch_flags & VMBUS_CHAN_FLAG_BATCHREAD) 392 task_fn = vmbus_chan_task; 393 else 394 task_fn = vmbus_chan_task_nobatch; 395 TASK_INIT(&chan->ch_task, 0, task_fn, chan); 396 397 /* TX bufring comes first */ 398 vmbus_txbr_setup(&chan->ch_txbr, br, txbr_size); 399 /* RX bufring immediately follows TX bufring */ 400 vmbus_rxbr_setup(&chan->ch_rxbr, br + txbr_size, rxbr_size); 401 402 /* Create sysctl tree for this channel */ 403 vmbus_chan_sysctl_create(chan); 404 405 /* 406 * Connect the bufrings, both RX and TX, to this channel. 407 */ 408 error = vmbus_chan_gpadl_connect(chan, cbr->cbr_paddr, 409 txbr_size + rxbr_size, &chan->ch_bufring_gpadl); 410 if (error) { 411 vmbus_chan_printf(chan, 412 "failed to connect bufring GPADL to chan%u\n", chan->ch_id); 413 goto failed; 414 } 415 416 /* 417 * Install this channel, before it is opened, but after everything 418 * else has been setup. 419 */ 420 vmbus_chan_set_chmap(chan); 421 422 /* 423 * Open channel w/ the bufring GPADL on the target CPU. 424 */ 425 mh = vmbus_msghc_get(sc, sizeof(*req)); 426 if (mh == NULL) { 427 vmbus_chan_printf(chan, 428 "can not get msg hypercall for chopen(chan%u)\n", 429 chan->ch_id); 430 error = ENXIO; 431 goto failed; 432 } 433 434 req = vmbus_msghc_dataptr(mh); 435 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHOPEN; 436 req->chm_chanid = chan->ch_id; 437 req->chm_openid = chan->ch_id; 438 req->chm_gpadl = chan->ch_bufring_gpadl; 439 req->chm_vcpuid = chan->ch_vcpuid; 440 req->chm_txbr_pgcnt = txbr_size >> PAGE_SHIFT; 441 if (udlen > 0) 442 memcpy(req->chm_udata, udata, udlen); 443 444 error = vmbus_msghc_exec(sc, mh); 445 if (error) { 446 vmbus_chan_printf(chan, 447 "chopen(chan%u) msg hypercall exec failed: %d\n", 448 chan->ch_id, error); 449 vmbus_msghc_put(sc, mh); 450 goto failed; 451 } 452 453 for (;;) { 454 msg = vmbus_msghc_poll_result(sc, mh); 455 if (msg != NULL) 456 break; 457 if (vmbus_chan_is_revoked(chan)) { 458 int i; 459 460 /* 461 * NOTE: 462 * Hypervisor does _not_ send response CHOPEN to 463 * a revoked channel. 464 */ 465 vmbus_chan_printf(chan, 466 "chan%u is revoked, when it is being opened\n", 467 chan->ch_id); 468 469 /* 470 * XXX 471 * Add extra delay before cancel the hypercall 472 * execution; mainly to close any possible 473 * CHRESCIND and CHOPEN_RESP races on the 474 * hypervisor side. 475 */ 476 #define REVOKE_LINGER 100 477 for (i = 0; i < REVOKE_LINGER; ++i) { 478 msg = vmbus_msghc_poll_result(sc, mh); 479 if (msg != NULL) 480 break; 481 pause("rchopen", 1); 482 } 483 #undef REVOKE_LINGER 484 if (msg == NULL) 485 vmbus_msghc_exec_cancel(sc, mh); 486 break; 487 } 488 pause("chopen", 1); 489 } 490 if (msg != NULL) { 491 status = ((const struct vmbus_chanmsg_chopen_resp *) 492 msg->msg_data)->chm_status; 493 } else { 494 /* XXX any non-0 value is ok here. */ 495 status = 0xff; 496 } 497 498 vmbus_msghc_put(sc, mh); 499 500 if (status == 0) { 501 if (bootverbose) 502 vmbus_chan_printf(chan, "chan%u opened\n", chan->ch_id); 503 return (0); 504 } 505 506 vmbus_chan_printf(chan, "failed to open chan%u\n", chan->ch_id); 507 error = ENXIO; 508 509 failed: 510 sysctl_ctx_free(&chan->ch_sysctl_ctx); 511 vmbus_chan_clear_chmap(chan); 512 if (chan->ch_bufring_gpadl != 0) { 513 int error1; 514 515 error1 = vmbus_chan_gpadl_disconnect(chan, 516 chan->ch_bufring_gpadl); 517 if (error1) { 518 /* 519 * Give caller a hint that the bufring GPADL is still 520 * connected. 521 */ 522 error = EISCONN; 523 } 524 chan->ch_bufring_gpadl = 0; 525 } 526 atomic_clear_int(&chan->ch_stflags, VMBUS_CHAN_ST_OPENED); 527 return (error); 528 } 529 530 int 531 vmbus_chan_gpadl_connect(struct vmbus_channel *chan, bus_addr_t paddr, 532 int size, uint32_t *gpadl0) 533 { 534 struct vmbus_softc *sc = chan->ch_vmbus; 535 struct vmbus_msghc *mh; 536 struct vmbus_chanmsg_gpadl_conn *req; 537 const struct vmbus_message *msg; 538 size_t reqsz; 539 uint32_t gpadl, status; 540 int page_count, range_len, i, cnt, error; 541 uint64_t page_id; 542 543 KASSERT(*gpadl0 == 0, ("GPADL is not zero")); 544 545 /* 546 * Preliminary checks. 547 */ 548 549 KASSERT((size & PAGE_MASK) == 0, 550 ("invalid GPA size %d, not multiple page size", size)); 551 page_count = size >> PAGE_SHIFT; 552 553 KASSERT((paddr & PAGE_MASK) == 0, 554 ("GPA is not page aligned %jx", (uintmax_t)paddr)); 555 page_id = paddr >> PAGE_SHIFT; 556 557 range_len = __offsetof(struct vmbus_gpa_range, gpa_page[page_count]); 558 /* 559 * We don't support multiple GPA ranges. 560 */ 561 if (range_len > UINT16_MAX) { 562 vmbus_chan_printf(chan, "GPA too large, %d pages\n", 563 page_count); 564 return EOPNOTSUPP; 565 } 566 567 /* 568 * Allocate GPADL id. 569 */ 570 gpadl = vmbus_gpadl_alloc(sc); 571 572 /* 573 * Connect this GPADL to the target channel. 574 * 575 * NOTE: 576 * Since each message can only hold small set of page 577 * addresses, several messages may be required to 578 * complete the connection. 579 */ 580 if (page_count > VMBUS_CHANMSG_GPADL_CONN_PGMAX) 581 cnt = VMBUS_CHANMSG_GPADL_CONN_PGMAX; 582 else 583 cnt = page_count; 584 page_count -= cnt; 585 586 reqsz = __offsetof(struct vmbus_chanmsg_gpadl_conn, 587 chm_range.gpa_page[cnt]); 588 mh = vmbus_msghc_get(sc, reqsz); 589 if (mh == NULL) { 590 vmbus_chan_printf(chan, 591 "can not get msg hypercall for gpadl_conn(chan%u)\n", 592 chan->ch_id); 593 return EIO; 594 } 595 596 req = vmbus_msghc_dataptr(mh); 597 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_CONN; 598 req->chm_chanid = chan->ch_id; 599 req->chm_gpadl = gpadl; 600 req->chm_range_len = range_len; 601 req->chm_range_cnt = 1; 602 req->chm_range.gpa_len = size; 603 req->chm_range.gpa_ofs = 0; 604 for (i = 0; i < cnt; ++i) 605 req->chm_range.gpa_page[i] = page_id++; 606 607 error = vmbus_msghc_exec(sc, mh); 608 if (error) { 609 vmbus_chan_printf(chan, 610 "gpadl_conn(chan%u) msg hypercall exec failed: %d\n", 611 chan->ch_id, error); 612 vmbus_msghc_put(sc, mh); 613 return error; 614 } 615 616 while (page_count > 0) { 617 struct vmbus_chanmsg_gpadl_subconn *subreq; 618 619 if (page_count > VMBUS_CHANMSG_GPADL_SUBCONN_PGMAX) 620 cnt = VMBUS_CHANMSG_GPADL_SUBCONN_PGMAX; 621 else 622 cnt = page_count; 623 page_count -= cnt; 624 625 reqsz = __offsetof(struct vmbus_chanmsg_gpadl_subconn, 626 chm_gpa_page[cnt]); 627 vmbus_msghc_reset(mh, reqsz); 628 629 subreq = vmbus_msghc_dataptr(mh); 630 subreq->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_SUBCONN; 631 subreq->chm_gpadl = gpadl; 632 for (i = 0; i < cnt; ++i) 633 subreq->chm_gpa_page[i] = page_id++; 634 635 vmbus_msghc_exec_noresult(mh); 636 } 637 KASSERT(page_count == 0, ("invalid page count %d", page_count)); 638 639 msg = vmbus_msghc_wait_result(sc, mh); 640 status = ((const struct vmbus_chanmsg_gpadl_connresp *) 641 msg->msg_data)->chm_status; 642 643 vmbus_msghc_put(sc, mh); 644 645 if (status != 0) { 646 vmbus_chan_printf(chan, "gpadl_conn(chan%u) failed: %u\n", 647 chan->ch_id, status); 648 return EIO; 649 } 650 651 /* Done; commit the GPADL id. */ 652 *gpadl0 = gpadl; 653 if (bootverbose) { 654 vmbus_chan_printf(chan, "gpadl_conn(chan%u) succeeded\n", 655 chan->ch_id); 656 } 657 return 0; 658 } 659 660 static bool 661 vmbus_chan_wait_revoke(const struct vmbus_channel *chan, bool can_sleep) 662 { 663 #define WAIT_COUNT 200 /* 200ms */ 664 665 int i; 666 667 for (i = 0; i < WAIT_COUNT; ++i) { 668 if (vmbus_chan_is_revoked(chan)) 669 return (true); 670 if (can_sleep) 671 pause("wchrev", 1); 672 else 673 DELAY(1000); 674 } 675 return (false); 676 677 #undef WAIT_COUNT 678 } 679 680 /* 681 * Disconnect the GPA from the target channel 682 */ 683 int 684 vmbus_chan_gpadl_disconnect(struct vmbus_channel *chan, uint32_t gpadl) 685 { 686 struct vmbus_softc *sc = chan->ch_vmbus; 687 struct vmbus_msghc *mh; 688 struct vmbus_chanmsg_gpadl_disconn *req; 689 int error; 690 691 KASSERT(gpadl != 0, ("GPADL is zero")); 692 693 mh = vmbus_msghc_get(sc, sizeof(*req)); 694 if (mh == NULL) { 695 vmbus_chan_printf(chan, 696 "can not get msg hypercall for gpadl_disconn(chan%u)\n", 697 chan->ch_id); 698 return (EBUSY); 699 } 700 701 req = vmbus_msghc_dataptr(mh); 702 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_DISCONN; 703 req->chm_chanid = chan->ch_id; 704 req->chm_gpadl = gpadl; 705 706 error = vmbus_msghc_exec(sc, mh); 707 if (error) { 708 vmbus_msghc_put(sc, mh); 709 710 if (vmbus_chan_wait_revoke(chan, true)) { 711 /* 712 * Error is benign; this channel is revoked, 713 * so this GPADL will not be touched anymore. 714 */ 715 vmbus_chan_printf(chan, 716 "gpadl_disconn(revoked chan%u) msg hypercall " 717 "exec failed: %d\n", chan->ch_id, error); 718 return (0); 719 } 720 vmbus_chan_printf(chan, 721 "gpadl_disconn(chan%u) msg hypercall exec failed: %d\n", 722 chan->ch_id, error); 723 return (error); 724 } 725 726 vmbus_msghc_wait_result(sc, mh); 727 /* Discard result; no useful information */ 728 vmbus_msghc_put(sc, mh); 729 730 return (0); 731 } 732 733 static void 734 vmbus_chan_detach(struct vmbus_channel *chan) 735 { 736 int refs; 737 738 KASSERT(chan->ch_refs > 0, ("chan%u: invalid refcnt %d", 739 chan->ch_id, chan->ch_refs)); 740 refs = atomic_fetchadd_int(&chan->ch_refs, -1); 741 #ifdef INVARIANTS 742 if (VMBUS_CHAN_ISPRIMARY(chan)) { 743 KASSERT(refs == 1, ("chan%u: invalid refcnt %d for prichan", 744 chan->ch_id, refs + 1)); 745 } 746 #endif 747 if (refs == 1) { 748 /* 749 * Detach the target channel. 750 */ 751 if (bootverbose) { 752 vmbus_chan_printf(chan, "chan%u detached\n", 753 chan->ch_id); 754 } 755 taskqueue_enqueue(chan->ch_mgmt_tq, &chan->ch_detach_task); 756 } 757 } 758 759 static void 760 vmbus_chan_clrchmap_task(void *xchan, int pending __unused) 761 { 762 struct vmbus_channel *chan = xchan; 763 764 critical_enter(); 765 chan->ch_vmbus->vmbus_chmap[chan->ch_id] = NULL; 766 critical_exit(); 767 } 768 769 static void 770 vmbus_chan_clear_chmap(struct vmbus_channel *chan) 771 { 772 struct task chmap_task; 773 774 TASK_INIT(&chmap_task, 0, vmbus_chan_clrchmap_task, chan); 775 vmbus_chan_run_task(chan, &chmap_task); 776 } 777 778 static void 779 vmbus_chan_set_chmap(struct vmbus_channel *chan) 780 { 781 __compiler_membar(); 782 chan->ch_vmbus->vmbus_chmap[chan->ch_id] = chan; 783 } 784 785 static int 786 vmbus_chan_close_internal(struct vmbus_channel *chan) 787 { 788 struct vmbus_softc *sc = chan->ch_vmbus; 789 struct vmbus_msghc *mh; 790 struct vmbus_chanmsg_chclose *req; 791 uint32_t old_stflags; 792 int error; 793 794 /* 795 * NOTE: 796 * Sub-channels are closed upon their primary channel closing, 797 * so they can be closed even before they are opened. 798 */ 799 for (;;) { 800 old_stflags = chan->ch_stflags; 801 if (atomic_cmpset_int(&chan->ch_stflags, old_stflags, 802 old_stflags & ~VMBUS_CHAN_ST_OPENED)) 803 break; 804 } 805 if ((old_stflags & VMBUS_CHAN_ST_OPENED) == 0) { 806 /* Not opened yet; done */ 807 if (bootverbose) { 808 vmbus_chan_printf(chan, "chan%u not opened\n", 809 chan->ch_id); 810 } 811 return (0); 812 } 813 814 /* 815 * Free this channel's sysctl tree attached to its device's 816 * sysctl tree. 817 */ 818 sysctl_ctx_free(&chan->ch_sysctl_ctx); 819 820 /* 821 * NOTE: 822 * Order is critical. This channel _must_ be uninstalled first, 823 * else the channel task may be enqueued by the IDT after it has 824 * been drained. 825 */ 826 vmbus_chan_clear_chmap(chan); 827 taskqueue_drain(chan->ch_tq, &chan->ch_task); 828 chan->ch_tq = NULL; 829 830 /* 831 * Close this channel. 832 */ 833 mh = vmbus_msghc_get(sc, sizeof(*req)); 834 if (mh == NULL) { 835 vmbus_chan_printf(chan, 836 "can not get msg hypercall for chclose(chan%u)\n", 837 chan->ch_id); 838 error = ENXIO; 839 goto disconnect; 840 } 841 842 req = vmbus_msghc_dataptr(mh); 843 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHCLOSE; 844 req->chm_chanid = chan->ch_id; 845 846 error = vmbus_msghc_exec_noresult(mh); 847 vmbus_msghc_put(sc, mh); 848 849 if (error) { 850 vmbus_chan_printf(chan, 851 "chclose(chan%u) msg hypercall exec failed: %d\n", 852 chan->ch_id, error); 853 goto disconnect; 854 } 855 856 if (bootverbose) 857 vmbus_chan_printf(chan, "chan%u closed\n", chan->ch_id); 858 859 disconnect: 860 /* 861 * Disconnect the TX+RX bufrings from this channel. 862 */ 863 if (chan->ch_bufring_gpadl != 0) { 864 int error1; 865 866 error1 = vmbus_chan_gpadl_disconnect(chan, 867 chan->ch_bufring_gpadl); 868 if (error1) { 869 /* 870 * XXX 871 * The bufring GPADL is still connected; abandon 872 * this bufring, instead of having mysterious 873 * crash or trashed data later on. 874 */ 875 vmbus_chan_printf(chan, "chan%u bufring GPADL " 876 "is still connected after close\n", chan->ch_id); 877 chan->ch_bufring = NULL; 878 /* 879 * Give caller a hint that the bufring GPADL is 880 * still connected. 881 */ 882 error = EISCONN; 883 } 884 chan->ch_bufring_gpadl = 0; 885 } 886 887 /* 888 * Destroy the TX+RX bufrings. 889 */ 890 if (chan->ch_bufring != NULL) { 891 hyperv_dmamem_free(&chan->ch_bufring_dma, chan->ch_bufring); 892 chan->ch_bufring = NULL; 893 } 894 return (error); 895 } 896 897 int 898 vmbus_chan_close_direct(struct vmbus_channel *chan) 899 { 900 int error; 901 902 #ifdef INVARIANTS 903 if (VMBUS_CHAN_ISPRIMARY(chan)) { 904 struct vmbus_channel *subchan; 905 906 /* 907 * All sub-channels _must_ have been closed, or are _not_ 908 * opened at all. 909 */ 910 mtx_lock(&chan->ch_subchan_lock); 911 TAILQ_FOREACH(subchan, &chan->ch_subchans, ch_sublink) { 912 KASSERT( 913 (subchan->ch_stflags & VMBUS_CHAN_ST_OPENED) == 0, 914 ("chan%u: subchan%u is still opened", 915 chan->ch_id, subchan->ch_subidx)); 916 } 917 mtx_unlock(&chan->ch_subchan_lock); 918 } 919 #endif 920 921 error = vmbus_chan_close_internal(chan); 922 if (!VMBUS_CHAN_ISPRIMARY(chan)) { 923 /* 924 * This sub-channel is referenced, when it is linked to 925 * the primary channel; drop that reference now. 926 */ 927 vmbus_chan_detach(chan); 928 } 929 return (error); 930 } 931 932 /* 933 * Caller should make sure that all sub-channels have 934 * been added to 'chan' and all to-be-closed channels 935 * are not being opened. 936 */ 937 void 938 vmbus_chan_close(struct vmbus_channel *chan) 939 { 940 int subchan_cnt; 941 942 if (!VMBUS_CHAN_ISPRIMARY(chan)) { 943 /* 944 * Sub-channel is closed when its primary channel 945 * is closed; done. 946 */ 947 return; 948 } 949 950 /* 951 * Close all sub-channels, if any. 952 */ 953 subchan_cnt = chan->ch_subchan_cnt; 954 if (subchan_cnt > 0) { 955 struct vmbus_channel **subchan; 956 int i; 957 958 subchan = vmbus_subchan_get(chan, subchan_cnt); 959 for (i = 0; i < subchan_cnt; ++i) { 960 vmbus_chan_close_internal(subchan[i]); 961 /* 962 * This sub-channel is referenced, when it is 963 * linked to the primary channel; drop that 964 * reference now. 965 */ 966 vmbus_chan_detach(subchan[i]); 967 } 968 vmbus_subchan_rel(subchan, subchan_cnt); 969 } 970 971 /* Then close the primary channel. */ 972 vmbus_chan_close_internal(chan); 973 } 974 975 void 976 vmbus_chan_intr_drain(struct vmbus_channel *chan) 977 { 978 979 taskqueue_drain(chan->ch_tq, &chan->ch_task); 980 } 981 982 int 983 vmbus_chan_send(struct vmbus_channel *chan, uint16_t type, uint16_t flags, 984 void *data, int dlen, uint64_t xactid) 985 { 986 struct vmbus_chanpkt pkt; 987 int pktlen, pad_pktlen, hlen, error; 988 uint64_t pad = 0; 989 struct iovec iov[3]; 990 boolean_t send_evt; 991 992 hlen = sizeof(pkt); 993 pktlen = hlen + dlen; 994 pad_pktlen = VMBUS_CHANPKT_TOTLEN(pktlen); 995 KASSERT(pad_pktlen <= vmbus_txbr_maxpktsz(&chan->ch_txbr), 996 ("invalid packet size %d", pad_pktlen)); 997 998 pkt.cp_hdr.cph_type = type; 999 pkt.cp_hdr.cph_flags = flags; 1000 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_hlen, hlen); 1001 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_tlen, pad_pktlen); 1002 pkt.cp_hdr.cph_xactid = xactid; 1003 1004 iov[0].iov_base = &pkt; 1005 iov[0].iov_len = hlen; 1006 iov[1].iov_base = data; 1007 iov[1].iov_len = dlen; 1008 iov[2].iov_base = &pad; 1009 iov[2].iov_len = pad_pktlen - pktlen; 1010 1011 error = vmbus_txbr_write(&chan->ch_txbr, iov, 3, &send_evt); 1012 if (!error && send_evt) 1013 vmbus_chan_signal_tx(chan); 1014 return error; 1015 } 1016 1017 int 1018 vmbus_chan_send_sglist(struct vmbus_channel *chan, 1019 struct vmbus_gpa sg[], int sglen, void *data, int dlen, uint64_t xactid) 1020 { 1021 struct vmbus_chanpkt_sglist pkt; 1022 int pktlen, pad_pktlen, hlen, error; 1023 struct iovec iov[4]; 1024 boolean_t send_evt; 1025 uint64_t pad = 0; 1026 1027 hlen = __offsetof(struct vmbus_chanpkt_sglist, cp_gpa[sglen]); 1028 pktlen = hlen + dlen; 1029 pad_pktlen = VMBUS_CHANPKT_TOTLEN(pktlen); 1030 KASSERT(pad_pktlen <= vmbus_txbr_maxpktsz(&chan->ch_txbr), 1031 ("invalid packet size %d", pad_pktlen)); 1032 1033 pkt.cp_hdr.cph_type = VMBUS_CHANPKT_TYPE_GPA; 1034 pkt.cp_hdr.cph_flags = VMBUS_CHANPKT_FLAG_RC; 1035 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_hlen, hlen); 1036 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_tlen, pad_pktlen); 1037 pkt.cp_hdr.cph_xactid = xactid; 1038 pkt.cp_rsvd = 0; 1039 pkt.cp_gpa_cnt = sglen; 1040 1041 iov[0].iov_base = &pkt; 1042 iov[0].iov_len = sizeof(pkt); 1043 iov[1].iov_base = sg; 1044 iov[1].iov_len = sizeof(struct vmbus_gpa) * sglen; 1045 iov[2].iov_base = data; 1046 iov[2].iov_len = dlen; 1047 iov[3].iov_base = &pad; 1048 iov[3].iov_len = pad_pktlen - pktlen; 1049 1050 error = vmbus_txbr_write(&chan->ch_txbr, iov, 4, &send_evt); 1051 if (!error && send_evt) 1052 vmbus_chan_signal_tx(chan); 1053 return error; 1054 } 1055 1056 int 1057 vmbus_chan_send_prplist(struct vmbus_channel *chan, 1058 struct vmbus_gpa_range *prp, int prp_cnt, void *data, int dlen, 1059 uint64_t xactid) 1060 { 1061 struct vmbus_chanpkt_prplist pkt; 1062 int pktlen, pad_pktlen, hlen, error; 1063 struct iovec iov[4]; 1064 boolean_t send_evt; 1065 uint64_t pad = 0; 1066 1067 hlen = __offsetof(struct vmbus_chanpkt_prplist, 1068 cp_range[0].gpa_page[prp_cnt]); 1069 pktlen = hlen + dlen; 1070 pad_pktlen = VMBUS_CHANPKT_TOTLEN(pktlen); 1071 KASSERT(pad_pktlen <= vmbus_txbr_maxpktsz(&chan->ch_txbr), 1072 ("invalid packet size %d", pad_pktlen)); 1073 1074 pkt.cp_hdr.cph_type = VMBUS_CHANPKT_TYPE_GPA; 1075 pkt.cp_hdr.cph_flags = VMBUS_CHANPKT_FLAG_RC; 1076 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_hlen, hlen); 1077 VMBUS_CHANPKT_SETLEN(pkt.cp_hdr.cph_tlen, pad_pktlen); 1078 pkt.cp_hdr.cph_xactid = xactid; 1079 pkt.cp_rsvd = 0; 1080 pkt.cp_range_cnt = 1; 1081 1082 iov[0].iov_base = &pkt; 1083 iov[0].iov_len = sizeof(pkt); 1084 iov[1].iov_base = prp; 1085 iov[1].iov_len = __offsetof(struct vmbus_gpa_range, gpa_page[prp_cnt]); 1086 iov[2].iov_base = data; 1087 iov[2].iov_len = dlen; 1088 iov[3].iov_base = &pad; 1089 iov[3].iov_len = pad_pktlen - pktlen; 1090 1091 error = vmbus_txbr_write(&chan->ch_txbr, iov, 4, &send_evt); 1092 if (!error && send_evt) 1093 vmbus_chan_signal_tx(chan); 1094 return error; 1095 } 1096 1097 int 1098 vmbus_chan_recv(struct vmbus_channel *chan, void *data, int *dlen0, 1099 uint64_t *xactid) 1100 { 1101 struct vmbus_chanpkt_hdr pkt; 1102 int error, dlen, hlen; 1103 1104 error = vmbus_rxbr_peek(&chan->ch_rxbr, &pkt, sizeof(pkt)); 1105 if (error) 1106 return (error); 1107 1108 if (__predict_false(pkt.cph_hlen < VMBUS_CHANPKT_HLEN_MIN)) { 1109 vmbus_chan_printf(chan, "invalid hlen %u\n", pkt.cph_hlen); 1110 /* XXX this channel is dead actually. */ 1111 return (EIO); 1112 } 1113 if (__predict_false(pkt.cph_hlen > pkt.cph_tlen)) { 1114 vmbus_chan_printf(chan, "invalid hlen %u and tlen %u\n", 1115 pkt.cph_hlen, pkt.cph_tlen); 1116 /* XXX this channel is dead actually. */ 1117 return (EIO); 1118 } 1119 1120 hlen = VMBUS_CHANPKT_GETLEN(pkt.cph_hlen); 1121 dlen = VMBUS_CHANPKT_GETLEN(pkt.cph_tlen) - hlen; 1122 1123 if (*dlen0 < dlen) { 1124 /* Return the size of this packet's data. */ 1125 *dlen0 = dlen; 1126 return (ENOBUFS); 1127 } 1128 1129 *xactid = pkt.cph_xactid; 1130 *dlen0 = dlen; 1131 1132 /* Skip packet header */ 1133 error = vmbus_rxbr_read(&chan->ch_rxbr, data, dlen, hlen); 1134 KASSERT(!error, ("vmbus_rxbr_read failed")); 1135 1136 return (0); 1137 } 1138 1139 int 1140 vmbus_chan_recv_pkt(struct vmbus_channel *chan, 1141 struct vmbus_chanpkt_hdr *pkt, int *pktlen0) 1142 { 1143 int error, pktlen, pkt_hlen; 1144 1145 pkt_hlen = sizeof(*pkt); 1146 error = vmbus_rxbr_peek(&chan->ch_rxbr, pkt, pkt_hlen); 1147 if (error) 1148 return (error); 1149 1150 if (__predict_false(pkt->cph_hlen < VMBUS_CHANPKT_HLEN_MIN)) { 1151 vmbus_chan_printf(chan, "invalid hlen %u\n", pkt->cph_hlen); 1152 /* XXX this channel is dead actually. */ 1153 return (EIO); 1154 } 1155 if (__predict_false(pkt->cph_hlen > pkt->cph_tlen)) { 1156 vmbus_chan_printf(chan, "invalid hlen %u and tlen %u\n", 1157 pkt->cph_hlen, pkt->cph_tlen); 1158 /* XXX this channel is dead actually. */ 1159 return (EIO); 1160 } 1161 1162 pktlen = VMBUS_CHANPKT_GETLEN(pkt->cph_tlen); 1163 if (*pktlen0 < pktlen) { 1164 /* Return the size of this packet. */ 1165 *pktlen0 = pktlen; 1166 return (ENOBUFS); 1167 } 1168 *pktlen0 = pktlen; 1169 1170 /* 1171 * Skip the fixed-size packet header, which has been filled 1172 * by the above vmbus_rxbr_peek(). 1173 */ 1174 error = vmbus_rxbr_read(&chan->ch_rxbr, pkt + 1, 1175 pktlen - pkt_hlen, pkt_hlen); 1176 KASSERT(!error, ("vmbus_rxbr_read failed")); 1177 1178 return (0); 1179 } 1180 1181 static void 1182 vmbus_chan_task(void *xchan, int pending __unused) 1183 { 1184 struct vmbus_channel *chan = xchan; 1185 vmbus_chan_callback_t cb = chan->ch_cb; 1186 void *cbarg = chan->ch_cbarg; 1187 1188 /* 1189 * Optimize host to guest signaling by ensuring: 1190 * 1. While reading the channel, we disable interrupts from 1191 * host. 1192 * 2. Ensure that we process all posted messages from the host 1193 * before returning from this callback. 1194 * 3. Once we return, enable signaling from the host. Once this 1195 * state is set we check to see if additional packets are 1196 * available to read. In this case we repeat the process. 1197 * 1198 * NOTE: Interrupt has been disabled in the ISR. 1199 */ 1200 for (;;) { 1201 uint32_t left; 1202 1203 cb(chan, cbarg); 1204 1205 left = vmbus_rxbr_intr_unmask(&chan->ch_rxbr); 1206 if (left == 0) { 1207 /* No more data in RX bufring; done */ 1208 break; 1209 } 1210 vmbus_rxbr_intr_mask(&chan->ch_rxbr); 1211 } 1212 } 1213 1214 static void 1215 vmbus_chan_task_nobatch(void *xchan, int pending __unused) 1216 { 1217 struct vmbus_channel *chan = xchan; 1218 1219 chan->ch_cb(chan, chan->ch_cbarg); 1220 } 1221 1222 static __inline void 1223 vmbus_event_flags_proc(struct vmbus_softc *sc, volatile u_long *event_flags, 1224 int flag_cnt) 1225 { 1226 int f; 1227 1228 for (f = 0; f < flag_cnt; ++f) { 1229 uint32_t chid_base; 1230 u_long flags; 1231 int chid_ofs; 1232 1233 if (event_flags[f] == 0) 1234 continue; 1235 1236 flags = atomic_swap_long(&event_flags[f], 0); 1237 chid_base = f << VMBUS_EVTFLAG_SHIFT; 1238 1239 while ((chid_ofs = ffsl(flags)) != 0) { 1240 struct vmbus_channel *chan; 1241 1242 --chid_ofs; /* NOTE: ffsl is 1-based */ 1243 flags &= ~(1UL << chid_ofs); 1244 1245 chan = sc->vmbus_chmap[chid_base + chid_ofs]; 1246 if (__predict_false(chan == NULL)) { 1247 /* Channel is closed. */ 1248 continue; 1249 } 1250 __compiler_membar(); 1251 1252 if (chan->ch_flags & VMBUS_CHAN_FLAG_BATCHREAD) 1253 vmbus_rxbr_intr_mask(&chan->ch_rxbr); 1254 taskqueue_enqueue(chan->ch_tq, &chan->ch_task); 1255 } 1256 } 1257 } 1258 1259 void 1260 vmbus_event_proc(struct vmbus_softc *sc, int cpu) 1261 { 1262 struct vmbus_evtflags *eventf; 1263 1264 /* 1265 * On Host with Win8 or above, the event page can be checked directly 1266 * to get the id of the channel that has the pending interrupt. 1267 */ 1268 eventf = VMBUS_PCPU_GET(sc, event_flags, cpu) + VMBUS_SINT_MESSAGE; 1269 vmbus_event_flags_proc(sc, eventf->evt_flags, 1270 VMBUS_PCPU_GET(sc, event_flags_cnt, cpu)); 1271 } 1272 1273 void 1274 vmbus_event_proc_compat(struct vmbus_softc *sc, int cpu) 1275 { 1276 struct vmbus_evtflags *eventf; 1277 1278 eventf = VMBUS_PCPU_GET(sc, event_flags, cpu) + VMBUS_SINT_MESSAGE; 1279 if (atomic_testandclear_long(&eventf->evt_flags[0], 0)) { 1280 vmbus_event_flags_proc(sc, sc->vmbus_rx_evtflags, 1281 VMBUS_CHAN_MAX_COMPAT >> VMBUS_EVTFLAG_SHIFT); 1282 } 1283 } 1284 1285 static void 1286 vmbus_chan_update_evtflagcnt(struct vmbus_softc *sc, 1287 const struct vmbus_channel *chan) 1288 { 1289 volatile int *flag_cnt_ptr; 1290 int flag_cnt; 1291 1292 flag_cnt = (chan->ch_id / VMBUS_EVTFLAG_LEN) + 1; 1293 flag_cnt_ptr = VMBUS_PCPU_PTR(sc, event_flags_cnt, chan->ch_cpuid); 1294 1295 for (;;) { 1296 int old_flag_cnt; 1297 1298 old_flag_cnt = *flag_cnt_ptr; 1299 if (old_flag_cnt >= flag_cnt) 1300 break; 1301 if (atomic_cmpset_int(flag_cnt_ptr, old_flag_cnt, flag_cnt)) { 1302 if (bootverbose) { 1303 vmbus_chan_printf(chan, 1304 "chan%u update cpu%d flag_cnt to %d\n", 1305 chan->ch_id, chan->ch_cpuid, flag_cnt); 1306 } 1307 break; 1308 } 1309 } 1310 } 1311 1312 static struct vmbus_channel * 1313 vmbus_chan_alloc(struct vmbus_softc *sc) 1314 { 1315 struct vmbus_channel *chan; 1316 1317 chan = malloc(sizeof(*chan), M_DEVBUF, M_WAITOK | M_ZERO); 1318 1319 chan->ch_monprm = hyperv_dmamem_alloc(bus_get_dma_tag(sc->vmbus_dev), 1320 HYPERCALL_PARAM_ALIGN, 0, sizeof(struct hyperv_mon_param), 1321 &chan->ch_monprm_dma, BUS_DMA_WAITOK | BUS_DMA_ZERO); 1322 if (chan->ch_monprm == NULL) { 1323 device_printf(sc->vmbus_dev, "monprm alloc failed\n"); 1324 free(chan, M_DEVBUF); 1325 return NULL; 1326 } 1327 1328 chan->ch_refs = 1; 1329 chan->ch_vmbus = sc; 1330 mtx_init(&chan->ch_subchan_lock, "vmbus subchan", NULL, MTX_DEF); 1331 sx_init(&chan->ch_orphan_lock, "vmbus chorphan"); 1332 TAILQ_INIT(&chan->ch_subchans); 1333 vmbus_rxbr_init(&chan->ch_rxbr); 1334 vmbus_txbr_init(&chan->ch_txbr); 1335 1336 return chan; 1337 } 1338 1339 static void 1340 vmbus_chan_free(struct vmbus_channel *chan) 1341 { 1342 1343 KASSERT(TAILQ_EMPTY(&chan->ch_subchans) && chan->ch_subchan_cnt == 0, 1344 ("still owns sub-channels")); 1345 KASSERT((chan->ch_stflags & 1346 (VMBUS_CHAN_ST_OPENED | 1347 VMBUS_CHAN_ST_ONPRIL | 1348 VMBUS_CHAN_ST_ONSUBL | 1349 VMBUS_CHAN_ST_ONLIST)) == 0, ("free busy channel")); 1350 KASSERT(chan->ch_orphan_xact == NULL, 1351 ("still has orphan xact installed")); 1352 KASSERT(chan->ch_refs == 0, ("chan%u: invalid refcnt %d", 1353 chan->ch_id, chan->ch_refs)); 1354 1355 hyperv_dmamem_free(&chan->ch_monprm_dma, chan->ch_monprm); 1356 mtx_destroy(&chan->ch_subchan_lock); 1357 sx_destroy(&chan->ch_orphan_lock); 1358 vmbus_rxbr_deinit(&chan->ch_rxbr); 1359 vmbus_txbr_deinit(&chan->ch_txbr); 1360 free(chan, M_DEVBUF); 1361 } 1362 1363 static int 1364 vmbus_chan_add(struct vmbus_channel *newchan) 1365 { 1366 struct vmbus_softc *sc = newchan->ch_vmbus; 1367 struct vmbus_channel *prichan; 1368 1369 if (newchan->ch_id == 0) { 1370 /* 1371 * XXX 1372 * Chan0 will neither be processed nor should be offered; 1373 * skip it. 1374 */ 1375 device_printf(sc->vmbus_dev, "got chan0 offer, discard\n"); 1376 return EINVAL; 1377 } else if (newchan->ch_id >= VMBUS_CHAN_MAX) { 1378 device_printf(sc->vmbus_dev, "invalid chan%u offer\n", 1379 newchan->ch_id); 1380 return EINVAL; 1381 } 1382 1383 mtx_lock(&sc->vmbus_prichan_lock); 1384 TAILQ_FOREACH(prichan, &sc->vmbus_prichans, ch_prilink) { 1385 /* 1386 * Sub-channel will have the same type GUID and instance 1387 * GUID as its primary channel. 1388 */ 1389 if (memcmp(&prichan->ch_guid_type, &newchan->ch_guid_type, 1390 sizeof(struct hyperv_guid)) == 0 && 1391 memcmp(&prichan->ch_guid_inst, &newchan->ch_guid_inst, 1392 sizeof(struct hyperv_guid)) == 0) 1393 break; 1394 } 1395 if (VMBUS_CHAN_ISPRIMARY(newchan)) { 1396 if (prichan == NULL) { 1397 /* Install the new primary channel */ 1398 vmbus_chan_ins_prilist(sc, newchan); 1399 mtx_unlock(&sc->vmbus_prichan_lock); 1400 goto done; 1401 } else { 1402 mtx_unlock(&sc->vmbus_prichan_lock); 1403 device_printf(sc->vmbus_dev, 1404 "duplicated primary chan%u\n", newchan->ch_id); 1405 return EINVAL; 1406 } 1407 } else { /* Sub-channel */ 1408 if (prichan == NULL) { 1409 mtx_unlock(&sc->vmbus_prichan_lock); 1410 device_printf(sc->vmbus_dev, 1411 "no primary chan for chan%u\n", newchan->ch_id); 1412 return EINVAL; 1413 } 1414 /* 1415 * Found the primary channel for this sub-channel and 1416 * move on. 1417 * 1418 * XXX refcnt prichan 1419 */ 1420 } 1421 mtx_unlock(&sc->vmbus_prichan_lock); 1422 1423 /* 1424 * This is a sub-channel; link it with the primary channel. 1425 */ 1426 KASSERT(!VMBUS_CHAN_ISPRIMARY(newchan), 1427 ("new channel is not sub-channel")); 1428 KASSERT(prichan != NULL, ("no primary channel")); 1429 1430 /* 1431 * Reference count this sub-channel; it will be dereferenced 1432 * when this sub-channel is closed. 1433 */ 1434 KASSERT(newchan->ch_refs == 1, ("chan%u: invalid refcnt %d", 1435 newchan->ch_id, newchan->ch_refs)); 1436 atomic_add_int(&newchan->ch_refs, 1); 1437 1438 newchan->ch_prichan = prichan; 1439 newchan->ch_dev = prichan->ch_dev; 1440 1441 mtx_lock(&prichan->ch_subchan_lock); 1442 vmbus_chan_ins_sublist(prichan, newchan); 1443 mtx_unlock(&prichan->ch_subchan_lock); 1444 /* 1445 * Notify anyone that is interested in this sub-channel, 1446 * after this sub-channel is setup. 1447 */ 1448 wakeup(prichan); 1449 done: 1450 /* 1451 * Hook this channel up for later revocation. 1452 */ 1453 mtx_lock(&sc->vmbus_chan_lock); 1454 vmbus_chan_ins_list(sc, newchan); 1455 mtx_unlock(&sc->vmbus_chan_lock); 1456 1457 if (bootverbose) { 1458 vmbus_chan_printf(newchan, "chan%u subidx%u offer\n", 1459 newchan->ch_id, newchan->ch_subidx); 1460 } 1461 1462 /* Select default cpu for this channel. */ 1463 vmbus_chan_cpu_default(newchan); 1464 1465 return 0; 1466 } 1467 1468 void 1469 vmbus_chan_cpu_set(struct vmbus_channel *chan, int cpu) 1470 { 1471 KASSERT(cpu >= 0 && cpu < mp_ncpus, ("invalid cpu %d", cpu)); 1472 1473 if (chan->ch_vmbus->vmbus_version == VMBUS_VERSION_WS2008 || 1474 chan->ch_vmbus->vmbus_version == VMBUS_VERSION_WIN7) { 1475 /* Only cpu0 is supported */ 1476 cpu = 0; 1477 } 1478 1479 chan->ch_cpuid = cpu; 1480 chan->ch_vcpuid = VMBUS_PCPU_GET(chan->ch_vmbus, vcpuid, cpu); 1481 1482 if (bootverbose) { 1483 vmbus_chan_printf(chan, 1484 "chan%u assigned to cpu%u [vcpu%u]\n", 1485 chan->ch_id, chan->ch_cpuid, chan->ch_vcpuid); 1486 } 1487 } 1488 1489 void 1490 vmbus_chan_cpu_rr(struct vmbus_channel *chan) 1491 { 1492 static uint32_t vmbus_chan_nextcpu; 1493 int cpu; 1494 1495 cpu = atomic_fetchadd_int(&vmbus_chan_nextcpu, 1) % mp_ncpus; 1496 vmbus_chan_cpu_set(chan, cpu); 1497 } 1498 1499 static void 1500 vmbus_chan_cpu_default(struct vmbus_channel *chan) 1501 { 1502 /* 1503 * By default, pin the channel to cpu0. Devices having 1504 * special channel-cpu mapping requirement should call 1505 * vmbus_chan_cpu_{set,rr}(). 1506 */ 1507 vmbus_chan_cpu_set(chan, 0); 1508 } 1509 1510 static void 1511 vmbus_chan_msgproc_choffer(struct vmbus_softc *sc, 1512 const struct vmbus_message *msg) 1513 { 1514 const struct vmbus_chanmsg_choffer *offer; 1515 struct vmbus_channel *chan; 1516 task_fn_t *detach_fn, *attach_fn; 1517 int error; 1518 1519 offer = (const struct vmbus_chanmsg_choffer *)msg->msg_data; 1520 1521 chan = vmbus_chan_alloc(sc); 1522 if (chan == NULL) { 1523 device_printf(sc->vmbus_dev, "allocate chan%u failed\n", 1524 offer->chm_chanid); 1525 return; 1526 } 1527 1528 chan->ch_id = offer->chm_chanid; 1529 chan->ch_subidx = offer->chm_subidx; 1530 chan->ch_guid_type = offer->chm_chtype; 1531 chan->ch_guid_inst = offer->chm_chinst; 1532 1533 /* Batch reading is on by default */ 1534 chan->ch_flags |= VMBUS_CHAN_FLAG_BATCHREAD; 1535 1536 chan->ch_monprm->mp_connid = VMBUS_CONNID_EVENT; 1537 if (sc->vmbus_version != VMBUS_VERSION_WS2008) 1538 chan->ch_monprm->mp_connid = offer->chm_connid; 1539 1540 if (offer->chm_flags1 & VMBUS_CHOFFER_FLAG1_HASMNF) { 1541 int trig_idx; 1542 1543 /* 1544 * Setup MNF stuffs. 1545 */ 1546 chan->ch_txflags |= VMBUS_CHAN_TXF_HASMNF; 1547 1548 trig_idx = offer->chm_montrig / VMBUS_MONTRIG_LEN; 1549 if (trig_idx >= VMBUS_MONTRIGS_MAX) 1550 panic("invalid monitor trigger %u", offer->chm_montrig); 1551 chan->ch_montrig = 1552 &sc->vmbus_mnf2->mnf_trigs[trig_idx].mt_pending; 1553 1554 chan->ch_montrig_mask = 1555 1 << (offer->chm_montrig % VMBUS_MONTRIG_LEN); 1556 } 1557 1558 /* 1559 * Setup event flag. 1560 */ 1561 chan->ch_evtflag = 1562 &sc->vmbus_tx_evtflags[chan->ch_id >> VMBUS_EVTFLAG_SHIFT]; 1563 chan->ch_evtflag_mask = 1UL << (chan->ch_id & VMBUS_EVTFLAG_MASK); 1564 1565 /* 1566 * Setup attach and detach tasks. 1567 */ 1568 if (VMBUS_CHAN_ISPRIMARY(chan)) { 1569 chan->ch_mgmt_tq = sc->vmbus_devtq; 1570 attach_fn = vmbus_prichan_attach_task; 1571 detach_fn = vmbus_prichan_detach_task; 1572 } else { 1573 chan->ch_mgmt_tq = sc->vmbus_subchtq; 1574 attach_fn = vmbus_subchan_attach_task; 1575 detach_fn = vmbus_subchan_detach_task; 1576 } 1577 TASK_INIT(&chan->ch_attach_task, 0, attach_fn, chan); 1578 TASK_INIT(&chan->ch_detach_task, 0, detach_fn, chan); 1579 1580 error = vmbus_chan_add(chan); 1581 if (error) { 1582 device_printf(sc->vmbus_dev, "add chan%u failed: %d\n", 1583 chan->ch_id, error); 1584 atomic_subtract_int(&chan->ch_refs, 1); 1585 vmbus_chan_free(chan); 1586 return; 1587 } 1588 taskqueue_enqueue(chan->ch_mgmt_tq, &chan->ch_attach_task); 1589 } 1590 1591 static void 1592 vmbus_chan_msgproc_chrescind(struct vmbus_softc *sc, 1593 const struct vmbus_message *msg) 1594 { 1595 const struct vmbus_chanmsg_chrescind *note; 1596 struct vmbus_channel *chan; 1597 1598 note = (const struct vmbus_chanmsg_chrescind *)msg->msg_data; 1599 if (note->chm_chanid > VMBUS_CHAN_MAX) { 1600 device_printf(sc->vmbus_dev, "invalid revoked chan%u\n", 1601 note->chm_chanid); 1602 return; 1603 } 1604 1605 /* 1606 * Find and remove the target channel from the channel list. 1607 */ 1608 mtx_lock(&sc->vmbus_chan_lock); 1609 TAILQ_FOREACH(chan, &sc->vmbus_chans, ch_link) { 1610 if (chan->ch_id == note->chm_chanid) 1611 break; 1612 } 1613 if (chan == NULL) { 1614 mtx_unlock(&sc->vmbus_chan_lock); 1615 device_printf(sc->vmbus_dev, "chan%u is not offered\n", 1616 note->chm_chanid); 1617 return; 1618 } 1619 vmbus_chan_rem_list(sc, chan); 1620 mtx_unlock(&sc->vmbus_chan_lock); 1621 1622 if (VMBUS_CHAN_ISPRIMARY(chan)) { 1623 /* 1624 * The target channel is a primary channel; remove the 1625 * target channel from the primary channel list now, 1626 * instead of later, so that it will not be found by 1627 * other sub-channel offers, which are processed in 1628 * this thread. 1629 */ 1630 mtx_lock(&sc->vmbus_prichan_lock); 1631 vmbus_chan_rem_prilist(sc, chan); 1632 mtx_unlock(&sc->vmbus_prichan_lock); 1633 } 1634 1635 /* 1636 * NOTE: 1637 * The following processing order is critical: 1638 * Set the REVOKED state flag before orphaning the installed xact. 1639 */ 1640 1641 if (atomic_testandset_int(&chan->ch_stflags, 1642 VMBUS_CHAN_ST_REVOKED_SHIFT)) 1643 panic("channel has already been revoked"); 1644 1645 sx_xlock(&chan->ch_orphan_lock); 1646 if (chan->ch_orphan_xact != NULL) 1647 vmbus_xact_ctx_orphan(chan->ch_orphan_xact); 1648 sx_xunlock(&chan->ch_orphan_lock); 1649 1650 if (bootverbose) 1651 vmbus_chan_printf(chan, "chan%u revoked\n", note->chm_chanid); 1652 vmbus_chan_detach(chan); 1653 } 1654 1655 static int 1656 vmbus_chan_release(struct vmbus_channel *chan) 1657 { 1658 struct vmbus_softc *sc = chan->ch_vmbus; 1659 struct vmbus_chanmsg_chfree *req; 1660 struct vmbus_msghc *mh; 1661 int error; 1662 1663 mh = vmbus_msghc_get(sc, sizeof(*req)); 1664 if (mh == NULL) { 1665 vmbus_chan_printf(chan, 1666 "can not get msg hypercall for chfree(chan%u)\n", 1667 chan->ch_id); 1668 return (ENXIO); 1669 } 1670 1671 req = vmbus_msghc_dataptr(mh); 1672 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHFREE; 1673 req->chm_chanid = chan->ch_id; 1674 1675 error = vmbus_msghc_exec_noresult(mh); 1676 vmbus_msghc_put(sc, mh); 1677 1678 if (error) { 1679 vmbus_chan_printf(chan, 1680 "chfree(chan%u) msg hypercall exec failed: %d\n", 1681 chan->ch_id, error); 1682 } else { 1683 if (bootverbose) 1684 vmbus_chan_printf(chan, "chan%u freed\n", chan->ch_id); 1685 } 1686 return (error); 1687 } 1688 1689 static void 1690 vmbus_prichan_detach_task(void *xchan, int pending __unused) 1691 { 1692 struct vmbus_channel *chan = xchan; 1693 1694 KASSERT(VMBUS_CHAN_ISPRIMARY(chan), 1695 ("chan%u is not primary channel", chan->ch_id)); 1696 1697 /* Delete and detach the device associated with this channel. */ 1698 vmbus_delete_child(chan); 1699 1700 /* Release this channel (back to vmbus). */ 1701 vmbus_chan_release(chan); 1702 1703 /* Free this channel's resource. */ 1704 vmbus_chan_free(chan); 1705 } 1706 1707 static void 1708 vmbus_subchan_detach_task(void *xchan, int pending __unused) 1709 { 1710 struct vmbus_channel *chan = xchan; 1711 struct vmbus_channel *pri_chan = chan->ch_prichan; 1712 1713 KASSERT(!VMBUS_CHAN_ISPRIMARY(chan), 1714 ("chan%u is primary channel", chan->ch_id)); 1715 1716 /* Release this channel (back to vmbus). */ 1717 vmbus_chan_release(chan); 1718 1719 /* Unlink from its primary channel's sub-channel list. */ 1720 mtx_lock(&pri_chan->ch_subchan_lock); 1721 vmbus_chan_rem_sublist(pri_chan, chan); 1722 mtx_unlock(&pri_chan->ch_subchan_lock); 1723 /* Notify anyone that is waiting for this sub-channel to vanish. */ 1724 wakeup(pri_chan); 1725 1726 /* Free this channel's resource. */ 1727 vmbus_chan_free(chan); 1728 } 1729 1730 static void 1731 vmbus_prichan_attach_task(void *xchan, int pending __unused) 1732 { 1733 1734 /* 1735 * Add device for this primary channel. 1736 */ 1737 vmbus_add_child(xchan); 1738 } 1739 1740 static void 1741 vmbus_subchan_attach_task(void *xchan __unused, int pending __unused) 1742 { 1743 1744 /* Nothing */ 1745 } 1746 1747 void 1748 vmbus_chan_destroy_all(struct vmbus_softc *sc) 1749 { 1750 1751 /* 1752 * Detach all devices and destroy the corresponding primary 1753 * channels. 1754 */ 1755 for (;;) { 1756 struct vmbus_channel *chan; 1757 1758 mtx_lock(&sc->vmbus_chan_lock); 1759 TAILQ_FOREACH(chan, &sc->vmbus_chans, ch_link) { 1760 if (VMBUS_CHAN_ISPRIMARY(chan)) 1761 break; 1762 } 1763 if (chan == NULL) { 1764 /* No more primary channels; done. */ 1765 mtx_unlock(&sc->vmbus_chan_lock); 1766 break; 1767 } 1768 vmbus_chan_rem_list(sc, chan); 1769 mtx_unlock(&sc->vmbus_chan_lock); 1770 1771 mtx_lock(&sc->vmbus_prichan_lock); 1772 vmbus_chan_rem_prilist(sc, chan); 1773 mtx_unlock(&sc->vmbus_prichan_lock); 1774 1775 taskqueue_enqueue(chan->ch_mgmt_tq, &chan->ch_detach_task); 1776 } 1777 } 1778 1779 struct vmbus_channel ** 1780 vmbus_subchan_get(struct vmbus_channel *pri_chan, int subchan_cnt) 1781 { 1782 struct vmbus_channel **ret, *chan; 1783 int i; 1784 1785 KASSERT(subchan_cnt > 0, ("invalid sub-channel count %d", subchan_cnt)); 1786 1787 ret = malloc(subchan_cnt * sizeof(struct vmbus_channel *), M_TEMP, 1788 M_WAITOK); 1789 1790 mtx_lock(&pri_chan->ch_subchan_lock); 1791 1792 while (pri_chan->ch_subchan_cnt < subchan_cnt) 1793 mtx_sleep(pri_chan, &pri_chan->ch_subchan_lock, 0, "subch", 0); 1794 1795 i = 0; 1796 TAILQ_FOREACH(chan, &pri_chan->ch_subchans, ch_sublink) { 1797 /* TODO: refcnt chan */ 1798 ret[i] = chan; 1799 1800 ++i; 1801 if (i == subchan_cnt) 1802 break; 1803 } 1804 KASSERT(i == subchan_cnt, ("invalid subchan count %d, should be %d", 1805 pri_chan->ch_subchan_cnt, subchan_cnt)); 1806 1807 mtx_unlock(&pri_chan->ch_subchan_lock); 1808 1809 return ret; 1810 } 1811 1812 void 1813 vmbus_subchan_rel(struct vmbus_channel **subchan, int subchan_cnt __unused) 1814 { 1815 1816 free(subchan, M_TEMP); 1817 } 1818 1819 void 1820 vmbus_subchan_drain(struct vmbus_channel *pri_chan) 1821 { 1822 mtx_lock(&pri_chan->ch_subchan_lock); 1823 while (pri_chan->ch_subchan_cnt > 0) 1824 mtx_sleep(pri_chan, &pri_chan->ch_subchan_lock, 0, "dsubch", 0); 1825 mtx_unlock(&pri_chan->ch_subchan_lock); 1826 } 1827 1828 void 1829 vmbus_chan_msgproc(struct vmbus_softc *sc, const struct vmbus_message *msg) 1830 { 1831 vmbus_chanmsg_proc_t msg_proc; 1832 uint32_t msg_type; 1833 1834 msg_type = ((const struct vmbus_chanmsg_hdr *)msg->msg_data)->chm_type; 1835 KASSERT(msg_type < VMBUS_CHANMSG_TYPE_MAX, 1836 ("invalid message type %u", msg_type)); 1837 1838 msg_proc = vmbus_chan_msgprocs[msg_type]; 1839 if (msg_proc != NULL) 1840 msg_proc(sc, msg); 1841 } 1842 1843 void 1844 vmbus_chan_set_readbatch(struct vmbus_channel *chan, bool on) 1845 { 1846 if (!on) 1847 chan->ch_flags &= ~VMBUS_CHAN_FLAG_BATCHREAD; 1848 else 1849 chan->ch_flags |= VMBUS_CHAN_FLAG_BATCHREAD; 1850 } 1851 1852 uint32_t 1853 vmbus_chan_id(const struct vmbus_channel *chan) 1854 { 1855 return chan->ch_id; 1856 } 1857 1858 uint32_t 1859 vmbus_chan_subidx(const struct vmbus_channel *chan) 1860 { 1861 return chan->ch_subidx; 1862 } 1863 1864 bool 1865 vmbus_chan_is_primary(const struct vmbus_channel *chan) 1866 { 1867 if (VMBUS_CHAN_ISPRIMARY(chan)) 1868 return true; 1869 else 1870 return false; 1871 } 1872 1873 const struct hyperv_guid * 1874 vmbus_chan_guid_inst(const struct vmbus_channel *chan) 1875 { 1876 return &chan->ch_guid_inst; 1877 } 1878 1879 int 1880 vmbus_chan_prplist_nelem(int br_size, int prpcnt_max, int dlen_max) 1881 { 1882 int elem_size; 1883 1884 elem_size = __offsetof(struct vmbus_chanpkt_prplist, 1885 cp_range[0].gpa_page[prpcnt_max]); 1886 elem_size += dlen_max; 1887 elem_size = VMBUS_CHANPKT_TOTLEN(elem_size); 1888 1889 return (vmbus_br_nelem(br_size, elem_size)); 1890 } 1891 1892 bool 1893 vmbus_chan_tx_empty(const struct vmbus_channel *chan) 1894 { 1895 1896 return (vmbus_txbr_empty(&chan->ch_txbr)); 1897 } 1898 1899 bool 1900 vmbus_chan_rx_empty(const struct vmbus_channel *chan) 1901 { 1902 1903 return (vmbus_rxbr_empty(&chan->ch_rxbr)); 1904 } 1905 1906 static int 1907 vmbus_chan_printf(const struct vmbus_channel *chan, const char *fmt, ...) 1908 { 1909 va_list ap; 1910 device_t dev; 1911 int retval; 1912 1913 if (chan->ch_dev == NULL || !device_is_alive(chan->ch_dev)) 1914 dev = chan->ch_vmbus->vmbus_dev; 1915 else 1916 dev = chan->ch_dev; 1917 1918 retval = device_print_prettyname(dev); 1919 va_start(ap, fmt); 1920 retval += vprintf(fmt, ap); 1921 va_end(ap); 1922 1923 return (retval); 1924 } 1925 1926 void 1927 vmbus_chan_run_task(struct vmbus_channel *chan, struct task *task) 1928 { 1929 1930 taskqueue_enqueue(chan->ch_tq, task); 1931 taskqueue_drain(chan->ch_tq, task); 1932 } 1933 1934 struct taskqueue * 1935 vmbus_chan_mgmt_tq(const struct vmbus_channel *chan) 1936 { 1937 1938 return (chan->ch_mgmt_tq); 1939 } 1940 1941 bool 1942 vmbus_chan_is_revoked(const struct vmbus_channel *chan) 1943 { 1944 1945 if (chan->ch_stflags & VMBUS_CHAN_ST_REVOKED) 1946 return (true); 1947 return (false); 1948 } 1949 1950 void 1951 vmbus_chan_set_orphan(struct vmbus_channel *chan, struct vmbus_xact_ctx *xact) 1952 { 1953 1954 sx_xlock(&chan->ch_orphan_lock); 1955 chan->ch_orphan_xact = xact; 1956 sx_xunlock(&chan->ch_orphan_lock); 1957 } 1958 1959 void 1960 vmbus_chan_unset_orphan(struct vmbus_channel *chan) 1961 { 1962 1963 sx_xlock(&chan->ch_orphan_lock); 1964 chan->ch_orphan_xact = NULL; 1965 sx_xunlock(&chan->ch_orphan_lock); 1966 } 1967 1968 const void * 1969 vmbus_chan_xact_wait(const struct vmbus_channel *chan, 1970 struct vmbus_xact *xact, size_t *resp_len, bool can_sleep) 1971 { 1972 const void *ret; 1973 1974 if (can_sleep) 1975 ret = vmbus_xact_wait(xact, resp_len); 1976 else 1977 ret = vmbus_xact_busywait(xact, resp_len); 1978 if (vmbus_chan_is_revoked(chan)) { 1979 /* 1980 * This xact probably is interrupted, and the 1981 * interruption can race the reply reception, 1982 * so we have to make sure that there are nothing 1983 * left on the RX bufring, i.e. this xact will 1984 * not be touched, once this function returns. 1985 * 1986 * Since the hypervisor will not put more data 1987 * onto the RX bufring once the channel is revoked, 1988 * the following loop will be terminated, once all 1989 * data are drained by the driver's channel 1990 * callback. 1991 */ 1992 while (!vmbus_chan_rx_empty(chan)) { 1993 if (can_sleep) 1994 pause("chxact", 1); 1995 else 1996 DELAY(1000); 1997 } 1998 } 1999 return (ret); 2000 } 2001