1 /*-
2 * Copyright (c) 2009-2012,2016-2017 Microsoft Corp.
3 * Copyright (c) 2012 NetApp Inc.
4 * Copyright (c) 2012 Citrix Inc.
5 * All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice unmodified, this list of conditions, and the following
12 * disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 */
28
29 /*
30 * VM Bus Driver Implementation
31 */
32
33 #include <sys/param.h>
34 #include <sys/bus.h>
35 #include <sys/kernel.h>
36 #include <sys/linker.h>
37 #include <sys/lock.h>
38 #include <sys/malloc.h>
39 #include <sys/module.h>
40 #include <sys/mutex.h>
41 #include <sys/sbuf.h>
42 #include <sys/smp.h>
43 #include <sys/sysctl.h>
44 #include <sys/systm.h>
45 #include <sys/taskqueue.h>
46
47 #include <vm/vm.h>
48 #include <vm/vm_extern.h>
49 #include <vm/vm_param.h>
50 #include <vm/pmap.h>
51
52 #include <machine/bus.h>
53 #if defined(__aarch64__)
54 #include <dev/psci/smccc.h>
55 #include <dev/hyperv/vmbus/aarch64/hyperv_machdep.h>
56 #include <dev/hyperv/vmbus/aarch64/hyperv_reg.h>
57 #else
58 #include <dev/hyperv/vmbus/x86/hyperv_machdep.h>
59 #include <dev/hyperv/vmbus/x86/hyperv_reg.h>
60 #include <machine/intr_machdep.h>
61 #include <x86/include/apicvar.h>
62 #endif
63 #include <machine/metadata.h>
64 #include <machine/md_var.h>
65 #include <machine/resource.h>
66 #include <contrib/dev/acpica/include/acpi.h>
67 #include <dev/acpica/acpivar.h>
68
69 #include <dev/hyperv/include/hyperv.h>
70 #include <dev/hyperv/include/vmbus_xact.h>
71 #include <dev/hyperv/vmbus/hyperv_var.h>
72 #include <dev/hyperv/vmbus/vmbus_reg.h>
73 #include <dev/hyperv/vmbus/vmbus_var.h>
74 #include <dev/hyperv/vmbus/vmbus_chanvar.h>
75 #include <dev/hyperv/vmbus/hyperv_common_reg.h>
76 #include "acpi_if.h"
77 #include "pcib_if.h"
78 #include "vmbus_if.h"
79
80 #define VMBUS_GPADL_START 0xe1e10
81
82 struct vmbus_msghc {
83 struct vmbus_xact *mh_xact;
84 struct hypercall_postmsg_in mh_inprm_save;
85 };
86
87 static void vmbus_identify(driver_t *, device_t);
88 static int vmbus_probe(device_t);
89 static int vmbus_attach(device_t);
90 static int vmbus_detach(device_t);
91 static int vmbus_read_ivar(device_t, device_t, int,
92 uintptr_t *);
93 static int vmbus_child_pnpinfo(device_t, device_t, struct sbuf *);
94 static struct resource *vmbus_alloc_resource(device_t dev,
95 device_t child, int type, int *rid,
96 rman_res_t start, rman_res_t end,
97 rman_res_t count, u_int flags);
98 static int vmbus_alloc_msi(device_t bus, device_t dev,
99 int count, int maxcount, int *irqs);
100 static int vmbus_release_msi(device_t bus, device_t dev,
101 int count, int *irqs);
102 static int vmbus_alloc_msix(device_t bus, device_t dev,
103 int *irq);
104 static int vmbus_release_msix(device_t bus, device_t dev,
105 int irq);
106 static int vmbus_map_msi(device_t bus, device_t dev,
107 int irq, uint64_t *addr, uint32_t *data);
108 static uint32_t vmbus_get_version_method(device_t, device_t);
109 static int vmbus_probe_guid_method(device_t, device_t,
110 const struct hyperv_guid *);
111 static uint32_t vmbus_get_vcpu_id_method(device_t bus,
112 device_t dev, int cpu);
113 static struct taskqueue *vmbus_get_eventtq_method(device_t, device_t,
114 int);
115 #if defined(EARLY_AP_STARTUP)
116 static void vmbus_intrhook(void *);
117 #endif
118
119 static int vmbus_init(struct vmbus_softc *);
120 static int vmbus_connect(struct vmbus_softc *, uint32_t);
121 static int vmbus_req_channels(struct vmbus_softc *sc);
122 static void vmbus_disconnect(struct vmbus_softc *);
123 static int vmbus_scan(struct vmbus_softc *);
124 static void vmbus_scan_teardown(struct vmbus_softc *);
125 static void vmbus_scan_done(struct vmbus_softc *,
126 const struct vmbus_message *);
127 static void vmbus_chanmsg_handle(struct vmbus_softc *,
128 const struct vmbus_message *);
129 static void vmbus_msg_task(void *, int);
130 static void vmbus_synic_setup(void *);
131 static void vmbus_synic_teardown(void *);
132 static int vmbus_sysctl_version(SYSCTL_HANDLER_ARGS);
133 static int vmbus_dma_alloc(struct vmbus_softc *);
134 static void vmbus_dma_free(struct vmbus_softc *);
135 static int vmbus_intr_setup(struct vmbus_softc *);
136 static void vmbus_intr_teardown(struct vmbus_softc *);
137 static int vmbus_doattach(struct vmbus_softc *);
138 static void vmbus_event_proc_dummy(struct vmbus_softc *,
139 int);
140 static bus_dma_tag_t vmbus_get_dma_tag(device_t parent, device_t child);
141 static struct vmbus_softc *vmbus_sc;
142 #if defined(__x86_64__)
143 static int vmbus_alloc_cpu_mem(struct vmbus_softc *sc);
144 static void vmbus_free_cpu_mem(struct vmbus_softc *sc);
145 #endif
146
147 SYSCTL_NODE(_hw, OID_AUTO, vmbus, CTLFLAG_RD | CTLFLAG_MPSAFE, NULL,
148 "Hyper-V vmbus");
149
150 static int vmbus_pin_evttask = 1;
151 SYSCTL_INT(_hw_vmbus, OID_AUTO, pin_evttask, CTLFLAG_RDTUN,
152 &vmbus_pin_evttask, 0, "Pin event tasks to their respective CPU");
153
154 #if defined(__x86_64__)
155 static int hv_tlb_hcall = 1;
156 SYSCTL_INT(_hw_vmbus, OID_AUTO, tlb_hcall , CTLFLAG_RDTUN,
157 &hv_tlb_hcall, 0, "Use Hyper_V hyercall for tlb flush");
158 #endif
159
160 uint32_t vmbus_current_version;
161
162 static const uint32_t vmbus_version[] = {
163 VMBUS_VERSION_WIN10,
164 VMBUS_VERSION_WIN8_1,
165 VMBUS_VERSION_WIN8,
166 VMBUS_VERSION_WIN7,
167 VMBUS_VERSION_WS2008
168 };
169
170 static const vmbus_chanmsg_proc_t
171 vmbus_chanmsg_handlers[VMBUS_CHANMSG_TYPE_MAX] = {
172 VMBUS_CHANMSG_PROC(CHOFFER_DONE, vmbus_scan_done),
173 VMBUS_CHANMSG_PROC_WAKEUP(CONNECT_RESP)
174 };
175
176 static device_method_t vmbus_methods[] = {
177 /* Device interface */
178 DEVMETHOD(device_identify, vmbus_identify),
179 DEVMETHOD(device_probe, vmbus_probe),
180 DEVMETHOD(device_attach, vmbus_attach),
181 DEVMETHOD(device_detach, vmbus_detach),
182 DEVMETHOD(device_shutdown, bus_generic_shutdown),
183 DEVMETHOD(device_suspend, bus_generic_suspend),
184 DEVMETHOD(device_resume, bus_generic_resume),
185
186 /* Bus interface */
187 DEVMETHOD(bus_add_child, bus_generic_add_child),
188 DEVMETHOD(bus_print_child, bus_generic_print_child),
189 DEVMETHOD(bus_read_ivar, vmbus_read_ivar),
190 DEVMETHOD(bus_child_pnpinfo, vmbus_child_pnpinfo),
191 DEVMETHOD(bus_alloc_resource, vmbus_alloc_resource),
192 DEVMETHOD(bus_release_resource, bus_generic_release_resource),
193 DEVMETHOD(bus_activate_resource, bus_generic_activate_resource),
194 DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource),
195 DEVMETHOD(bus_setup_intr, bus_generic_setup_intr),
196 DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr),
197 DEVMETHOD(bus_get_cpus, bus_generic_get_cpus),
198 DEVMETHOD(bus_get_dma_tag, vmbus_get_dma_tag),
199
200 /* pcib interface */
201 DEVMETHOD(pcib_alloc_msi, vmbus_alloc_msi),
202 DEVMETHOD(pcib_release_msi, vmbus_release_msi),
203 DEVMETHOD(pcib_alloc_msix, vmbus_alloc_msix),
204 DEVMETHOD(pcib_release_msix, vmbus_release_msix),
205 DEVMETHOD(pcib_map_msi, vmbus_map_msi),
206
207 /* Vmbus interface */
208 DEVMETHOD(vmbus_get_version, vmbus_get_version_method),
209 DEVMETHOD(vmbus_probe_guid, vmbus_probe_guid_method),
210 DEVMETHOD(vmbus_get_vcpu_id, vmbus_get_vcpu_id_method),
211 DEVMETHOD(vmbus_get_event_taskq, vmbus_get_eventtq_method),
212
213 DEVMETHOD_END
214 };
215
216 static driver_t vmbus_driver = {
217 "vmbus",
218 vmbus_methods,
219 sizeof(struct vmbus_softc)
220 };
221
222 uint32_t hv_max_vp_index;
223
224 DRIVER_MODULE(vmbus, pcib, vmbus_driver, NULL, NULL);
225 DRIVER_MODULE(vmbus, acpi_syscontainer, vmbus_driver, NULL, NULL);
226
227 MODULE_DEPEND(vmbus, acpi, 1, 1, 1);
228 MODULE_DEPEND(vmbus, pci, 1, 1, 1);
229 MODULE_VERSION(vmbus, 1);
230
231 static __inline struct vmbus_softc *
vmbus_get_softc(void)232 vmbus_get_softc(void)
233 {
234 return vmbus_sc;
235 }
236
237 static bus_dma_tag_t
vmbus_get_dma_tag(device_t dev,device_t child)238 vmbus_get_dma_tag(device_t dev, device_t child)
239 {
240 struct vmbus_softc *sc = vmbus_get_softc();
241 return (sc->dmat);
242 }
243
244 void
vmbus_msghc_reset(struct vmbus_msghc * mh,size_t dsize)245 vmbus_msghc_reset(struct vmbus_msghc *mh, size_t dsize)
246 {
247 struct hypercall_postmsg_in *inprm;
248
249 if (dsize > HYPERCALL_POSTMSGIN_DSIZE_MAX)
250 panic("invalid data size %zu", dsize);
251
252 inprm = vmbus_xact_req_data(mh->mh_xact);
253 memset(inprm, 0, HYPERCALL_POSTMSGIN_SIZE);
254 inprm->hc_connid = VMBUS_CONNID_MESSAGE;
255 inprm->hc_msgtype = HYPERV_MSGTYPE_CHANNEL;
256 inprm->hc_dsize = dsize;
257 }
258
259 struct vmbus_msghc *
vmbus_msghc_get(struct vmbus_softc * sc,size_t dsize)260 vmbus_msghc_get(struct vmbus_softc *sc, size_t dsize)
261 {
262 struct vmbus_msghc *mh;
263 struct vmbus_xact *xact;
264
265 if (dsize > HYPERCALL_POSTMSGIN_DSIZE_MAX)
266 panic("invalid data size %zu", dsize);
267
268 xact = vmbus_xact_get(sc->vmbus_xc,
269 dsize + __offsetof(struct hypercall_postmsg_in, hc_data[0]));
270 if (xact == NULL)
271 return (NULL);
272
273 mh = vmbus_xact_priv(xact, sizeof(*mh));
274 mh->mh_xact = xact;
275
276 vmbus_msghc_reset(mh, dsize);
277 return (mh);
278 }
279
280 void
vmbus_msghc_put(struct vmbus_softc * sc __unused,struct vmbus_msghc * mh)281 vmbus_msghc_put(struct vmbus_softc *sc __unused, struct vmbus_msghc *mh)
282 {
283
284 vmbus_xact_put(mh->mh_xact);
285 }
286
287 void *
vmbus_msghc_dataptr(struct vmbus_msghc * mh)288 vmbus_msghc_dataptr(struct vmbus_msghc *mh)
289 {
290 struct hypercall_postmsg_in *inprm;
291
292 inprm = vmbus_xact_req_data(mh->mh_xact);
293 return (inprm->hc_data);
294 }
295
296 int
vmbus_msghc_exec_noresult(struct vmbus_msghc * mh)297 vmbus_msghc_exec_noresult(struct vmbus_msghc *mh)
298 {
299 sbintime_t time = SBT_1MS;
300 struct hypercall_postmsg_in *inprm;
301 bus_addr_t inprm_paddr;
302 int i;
303
304 inprm = vmbus_xact_req_data(mh->mh_xact);
305 inprm_paddr = vmbus_xact_req_paddr(mh->mh_xact);
306
307 /*
308 * Save the input parameter so that we could restore the input
309 * parameter if the Hypercall failed.
310 *
311 * XXX
312 * Is this really necessary?! i.e. Will the Hypercall ever
313 * overwrite the input parameter?
314 */
315 memcpy(&mh->mh_inprm_save, inprm, HYPERCALL_POSTMSGIN_SIZE);
316
317 /*
318 * In order to cope with transient failures, e.g. insufficient
319 * resources on host side, we retry the post message Hypercall
320 * several times. 20 retries seem sufficient.
321 */
322 #define HC_RETRY_MAX 20
323
324 for (i = 0; i < HC_RETRY_MAX; ++i) {
325 uint64_t status;
326
327 status = hypercall_post_message(inprm_paddr);
328 if (status == HYPERCALL_STATUS_SUCCESS)
329 return 0;
330
331 pause_sbt("hcpmsg", time, 0, C_HARDCLOCK);
332 if (time < SBT_1S * 2)
333 time *= 2;
334
335 /* Restore input parameter and try again */
336 memcpy(inprm, &mh->mh_inprm_save, HYPERCALL_POSTMSGIN_SIZE);
337 }
338
339 #undef HC_RETRY_MAX
340
341 return EIO;
342 }
343
344 int
vmbus_msghc_exec(struct vmbus_softc * sc __unused,struct vmbus_msghc * mh)345 vmbus_msghc_exec(struct vmbus_softc *sc __unused, struct vmbus_msghc *mh)
346 {
347 int error;
348
349 vmbus_xact_activate(mh->mh_xact);
350 error = vmbus_msghc_exec_noresult(mh);
351 if (error)
352 vmbus_xact_deactivate(mh->mh_xact);
353 return error;
354 }
355
356 void
vmbus_msghc_exec_cancel(struct vmbus_softc * sc __unused,struct vmbus_msghc * mh)357 vmbus_msghc_exec_cancel(struct vmbus_softc *sc __unused, struct vmbus_msghc *mh)
358 {
359
360 vmbus_xact_deactivate(mh->mh_xact);
361 }
362
363 const struct vmbus_message *
vmbus_msghc_wait_result(struct vmbus_softc * sc __unused,struct vmbus_msghc * mh)364 vmbus_msghc_wait_result(struct vmbus_softc *sc __unused, struct vmbus_msghc *mh)
365 {
366 size_t resp_len;
367
368 return (vmbus_xact_wait(mh->mh_xact, &resp_len));
369 }
370
371 const struct vmbus_message *
vmbus_msghc_poll_result(struct vmbus_softc * sc __unused,struct vmbus_msghc * mh)372 vmbus_msghc_poll_result(struct vmbus_softc *sc __unused, struct vmbus_msghc *mh)
373 {
374 size_t resp_len;
375
376 return (vmbus_xact_poll(mh->mh_xact, &resp_len));
377 }
378
379 void
vmbus_msghc_wakeup(struct vmbus_softc * sc,const struct vmbus_message * msg)380 vmbus_msghc_wakeup(struct vmbus_softc *sc, const struct vmbus_message *msg)
381 {
382
383 vmbus_xact_ctx_wakeup(sc->vmbus_xc, msg, sizeof(*msg));
384 }
385
386 uint32_t
vmbus_gpadl_alloc(struct vmbus_softc * sc)387 vmbus_gpadl_alloc(struct vmbus_softc *sc)
388 {
389 uint32_t gpadl;
390
391 again:
392 gpadl = atomic_fetchadd_int(&sc->vmbus_gpadl, 1);
393 if (gpadl == 0)
394 goto again;
395 return (gpadl);
396 }
397
398 /* Used for Hyper-V socket when guest client connects to host */
399 int
vmbus_req_tl_connect(struct hyperv_guid * guest_srv_id,struct hyperv_guid * host_srv_id)400 vmbus_req_tl_connect(struct hyperv_guid *guest_srv_id,
401 struct hyperv_guid *host_srv_id)
402 {
403 struct vmbus_softc *sc = vmbus_get_softc();
404 struct vmbus_chanmsg_tl_connect *req;
405 struct vmbus_msghc *mh;
406 int error;
407
408 if (!sc)
409 return ENXIO;
410
411 mh = vmbus_msghc_get(sc, sizeof(*req));
412 if (mh == NULL) {
413 device_printf(sc->vmbus_dev,
414 "can not get msg hypercall for tl connect\n");
415 return ENXIO;
416 }
417
418 req = vmbus_msghc_dataptr(mh);
419 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_TL_CONN;
420 req->guest_endpoint_id = *guest_srv_id;
421 req->host_service_id = *host_srv_id;
422
423 error = vmbus_msghc_exec_noresult(mh);
424 vmbus_msghc_put(sc, mh);
425
426 if (error) {
427 device_printf(sc->vmbus_dev,
428 "tl connect msg hypercall failed\n");
429 }
430
431 return error;
432 }
433
434 static int
vmbus_connect(struct vmbus_softc * sc,uint32_t version)435 vmbus_connect(struct vmbus_softc *sc, uint32_t version)
436 {
437 struct vmbus_chanmsg_connect *req;
438 const struct vmbus_message *msg;
439 struct vmbus_msghc *mh;
440 int error, done = 0;
441
442 mh = vmbus_msghc_get(sc, sizeof(*req));
443 if (mh == NULL)
444 return ENXIO;
445
446 req = vmbus_msghc_dataptr(mh);
447 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CONNECT;
448 req->chm_ver = version;
449 req->chm_evtflags = pmap_kextract((vm_offset_t)sc->vmbus_evtflags);
450 req->chm_mnf1 = pmap_kextract((vm_offset_t)sc->vmbus_mnf1);
451 req->chm_mnf2 = pmap_kextract((vm_offset_t)sc->vmbus_mnf2);
452
453 error = vmbus_msghc_exec(sc, mh);
454 if (error) {
455 vmbus_msghc_put(sc, mh);
456 return error;
457 }
458
459 msg = vmbus_msghc_wait_result(sc, mh);
460 done = ((const struct vmbus_chanmsg_connect_resp *)
461 msg->msg_data)->chm_done;
462
463 vmbus_msghc_put(sc, mh);
464
465 return (done ? 0 : EOPNOTSUPP);
466 }
467
468 static int
vmbus_init(struct vmbus_softc * sc)469 vmbus_init(struct vmbus_softc *sc)
470 {
471 int i;
472
473 for (i = 0; i < nitems(vmbus_version); ++i) {
474 int error;
475
476 error = vmbus_connect(sc, vmbus_version[i]);
477 if (!error) {
478 vmbus_current_version = vmbus_version[i];
479 sc->vmbus_version = vmbus_version[i];
480 device_printf(sc->vmbus_dev, "version %u.%u\n",
481 VMBUS_VERSION_MAJOR(sc->vmbus_version),
482 VMBUS_VERSION_MINOR(sc->vmbus_version));
483 return 0;
484 }
485 }
486 return ENXIO;
487 }
488
489 static void
vmbus_disconnect(struct vmbus_softc * sc)490 vmbus_disconnect(struct vmbus_softc *sc)
491 {
492 struct vmbus_chanmsg_disconnect *req;
493 struct vmbus_msghc *mh;
494 int error;
495
496 mh = vmbus_msghc_get(sc, sizeof(*req));
497 if (mh == NULL) {
498 device_printf(sc->vmbus_dev,
499 "can not get msg hypercall for disconnect\n");
500 return;
501 }
502
503 req = vmbus_msghc_dataptr(mh);
504 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_DISCONNECT;
505
506 error = vmbus_msghc_exec_noresult(mh);
507 vmbus_msghc_put(sc, mh);
508
509 if (error) {
510 device_printf(sc->vmbus_dev,
511 "disconnect msg hypercall failed\n");
512 }
513 }
514
515 static int
vmbus_req_channels(struct vmbus_softc * sc)516 vmbus_req_channels(struct vmbus_softc *sc)
517 {
518 struct vmbus_chanmsg_chrequest *req;
519 struct vmbus_msghc *mh;
520 int error;
521
522 mh = vmbus_msghc_get(sc, sizeof(*req));
523 if (mh == NULL)
524 return ENXIO;
525
526 req = vmbus_msghc_dataptr(mh);
527 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHREQUEST;
528
529 error = vmbus_msghc_exec_noresult(mh);
530 vmbus_msghc_put(sc, mh);
531
532 return error;
533 }
534
535 static void
vmbus_scan_done_task(void * xsc,int pending __unused)536 vmbus_scan_done_task(void *xsc, int pending __unused)
537 {
538 struct vmbus_softc *sc = xsc;
539
540 bus_topo_lock();
541 sc->vmbus_scandone = true;
542 bus_topo_unlock();
543 wakeup(&sc->vmbus_scandone);
544 }
545
546 static void
vmbus_scan_done(struct vmbus_softc * sc,const struct vmbus_message * msg __unused)547 vmbus_scan_done(struct vmbus_softc *sc,
548 const struct vmbus_message *msg __unused)
549 {
550
551 taskqueue_enqueue(sc->vmbus_devtq, &sc->vmbus_scandone_task);
552 }
553
554 static int
vmbus_scan(struct vmbus_softc * sc)555 vmbus_scan(struct vmbus_softc *sc)
556 {
557 int error;
558
559 /*
560 * Identify, probe and attach for non-channel devices.
561 */
562 bus_identify_children(sc->vmbus_dev);
563 bus_attach_children(sc->vmbus_dev);
564
565 /*
566 * This taskqueue serializes vmbus devices' attach and detach
567 * for channel offer and rescind messages.
568 */
569 sc->vmbus_devtq = taskqueue_create("vmbus dev", M_WAITOK,
570 taskqueue_thread_enqueue, &sc->vmbus_devtq);
571 taskqueue_start_threads(&sc->vmbus_devtq, 1, PI_NET, "vmbusdev");
572 TASK_INIT(&sc->vmbus_scandone_task, 0, vmbus_scan_done_task, sc);
573
574 /*
575 * This taskqueue handles sub-channel detach, so that vmbus
576 * device's detach running in vmbus_devtq can drain its sub-
577 * channels.
578 */
579 sc->vmbus_subchtq = taskqueue_create("vmbus subch", M_WAITOK,
580 taskqueue_thread_enqueue, &sc->vmbus_subchtq);
581 taskqueue_start_threads(&sc->vmbus_subchtq, 1, PI_NET, "vmbussch");
582
583 /*
584 * Start vmbus scanning.
585 */
586 error = vmbus_req_channels(sc);
587 if (error) {
588 device_printf(sc->vmbus_dev, "channel request failed: %d\n",
589 error);
590 return (error);
591 }
592
593 /*
594 * Wait for all vmbus devices from the initial channel offers to be
595 * attached.
596 */
597 bus_topo_assert();
598 while (!sc->vmbus_scandone)
599 mtx_sleep(&sc->vmbus_scandone, bus_topo_mtx(), 0, "vmbusdev", 0);
600
601 if (bootverbose) {
602 device_printf(sc->vmbus_dev, "device scan, probe and attach "
603 "done\n");
604 }
605 return (0);
606 }
607
608 static void
vmbus_scan_teardown(struct vmbus_softc * sc)609 vmbus_scan_teardown(struct vmbus_softc *sc)
610 {
611
612 bus_topo_assert();
613 if (sc->vmbus_devtq != NULL) {
614 bus_topo_unlock();
615 taskqueue_free(sc->vmbus_devtq);
616 bus_topo_lock();
617 sc->vmbus_devtq = NULL;
618 }
619 if (sc->vmbus_subchtq != NULL) {
620 bus_topo_unlock();
621 taskqueue_free(sc->vmbus_subchtq);
622 bus_topo_lock();
623 sc->vmbus_subchtq = NULL;
624 }
625 }
626
627 static void
vmbus_chanmsg_handle(struct vmbus_softc * sc,const struct vmbus_message * msg)628 vmbus_chanmsg_handle(struct vmbus_softc *sc, const struct vmbus_message *msg)
629 {
630 vmbus_chanmsg_proc_t msg_proc;
631 uint32_t msg_type;
632
633 msg_type = ((const struct vmbus_chanmsg_hdr *)msg->msg_data)->chm_type;
634 if (msg_type >= VMBUS_CHANMSG_TYPE_MAX) {
635 device_printf(sc->vmbus_dev, "unknown message type 0x%x\n",
636 msg_type);
637 return;
638 }
639
640 msg_proc = vmbus_chanmsg_handlers[msg_type];
641 if (msg_proc != NULL)
642 msg_proc(sc, msg);
643
644 /* Channel specific processing */
645 vmbus_chan_msgproc(sc, msg);
646 }
647
648 static void
vmbus_msg_task(void * xsc,int pending __unused)649 vmbus_msg_task(void *xsc, int pending __unused)
650 {
651 struct vmbus_softc *sc = xsc;
652 volatile struct vmbus_message *msg;
653
654 msg = VMBUS_PCPU_GET(sc, message, curcpu) + VMBUS_SINT_MESSAGE;
655 for (;;) {
656 if (msg->msg_type == HYPERV_MSGTYPE_NONE) {
657 /* No message */
658 break;
659 } else if (msg->msg_type == HYPERV_MSGTYPE_CHANNEL) {
660 /* Channel message */
661 vmbus_chanmsg_handle(sc,
662 __DEVOLATILE(const struct vmbus_message *, msg));
663 }
664
665 msg->msg_type = HYPERV_MSGTYPE_NONE;
666 /*
667 * Make sure the write to msg_type (i.e. set to
668 * HYPERV_MSGTYPE_NONE) happens before we read the
669 * msg_flags and EOMing. Otherwise, the EOMing will
670 * not deliver any more messages since there is no
671 * empty slot
672 *
673 * NOTE:
674 * mb() is used here, since atomic_thread_fence_seq_cst()
675 * will become compiler fence on UP kernel.
676 */
677 mb();
678 if (msg->msg_flags & VMBUS_MSGFLAG_PENDING) {
679 /*
680 * This will cause message queue rescan to possibly
681 * deliver another msg from the hypervisor
682 */
683 WRMSR(MSR_HV_EOM, 0);
684 }
685 }
686 }
687 static __inline int
vmbus_handle_intr1(struct vmbus_softc * sc,struct trapframe * frame,int cpu)688 vmbus_handle_intr1(struct vmbus_softc *sc, struct trapframe *frame, int cpu)
689 {
690 volatile struct vmbus_message *msg;
691 struct vmbus_message *msg_base;
692
693 msg_base = VMBUS_PCPU_GET(sc, message, cpu);
694
695 /*
696 * Check event timer.
697 *
698 * TODO: move this to independent IDT vector.
699 */
700 vmbus_handle_timer_intr1(msg_base, frame);
701 /*
702 * Check events. Hot path for network and storage I/O data; high rate.
703 *
704 * NOTE:
705 * As recommended by the Windows guest fellows, we check events before
706 * checking messages.
707 */
708 sc->vmbus_event_proc(sc, cpu);
709
710 /*
711 * Check messages. Mainly management stuffs; ultra low rate.
712 */
713 msg = msg_base + VMBUS_SINT_MESSAGE;
714 if (__predict_false(msg->msg_type != HYPERV_MSGTYPE_NONE)) {
715 taskqueue_enqueue(VMBUS_PCPU_GET(sc, message_tq, cpu),
716 VMBUS_PCPU_PTR(sc, message_task, cpu));
717 }
718
719 return (FILTER_HANDLED);
720 }
721
722 void
vmbus_handle_intr(struct trapframe * trap_frame)723 vmbus_handle_intr(struct trapframe *trap_frame)
724 {
725 struct vmbus_softc *sc = vmbus_get_softc();
726 int cpu = curcpu;
727
728 /*
729 * Disable preemption.
730 */
731 critical_enter();
732
733 /*
734 * Do a little interrupt counting. This used x86 specific
735 * intrcnt_add function
736 */
737 #if !defined(__aarch64__)
738 (*VMBUS_PCPU_GET(sc, intr_cnt, cpu))++;
739 #endif /* not for aarch64 */
740 vmbus_handle_intr1(sc, trap_frame, cpu);
741
742 /*
743 * Enable preemption.
744 */
745 critical_exit();
746 }
747
748 static void
vmbus_synic_setup(void * xsc)749 vmbus_synic_setup(void *xsc)
750 {
751 struct vmbus_softc *sc = xsc;
752 int cpu = curcpu;
753 uint64_t val, orig;
754 uint32_t sint;
755
756 if (hyperv_features & CPUID_HV_MSR_VP_INDEX) {
757 /* Save virtual processor id. */
758 VMBUS_PCPU_GET(sc, vcpuid, cpu) = RDMSR(MSR_HV_VP_INDEX);
759 } else {
760 /* Set virtual processor id to 0 for compatibility. */
761 VMBUS_PCPU_GET(sc, vcpuid, cpu) = 0;
762 }
763
764 if (VMBUS_PCPU_GET(sc, vcpuid, cpu) > hv_max_vp_index)
765 hv_max_vp_index = VMBUS_PCPU_GET(sc, vcpuid, cpu);
766
767 /*
768 * Setup the SynIC message.
769 */
770 orig = RDMSR(MSR_HV_SIMP);
771 val = pmap_kextract((vm_offset_t)VMBUS_PCPU_GET(sc, message, cpu)) &
772 MSR_HV_SIMP_PGMASK;
773 val |= MSR_HV_SIMP_ENABLE | (orig & MSR_HV_SIMP_RSVD_MASK);
774 WRMSR(MSR_HV_SIMP, val);
775 /*
776 * Setup the SynIC event flags.
777 */
778 orig = RDMSR(MSR_HV_SIEFP);
779 val = pmap_kextract((vm_offset_t)VMBUS_PCPU_GET(sc, event_flags, cpu)) &
780 MSR_HV_SIMP_PGMASK;
781 val |= MSR_HV_SIEFP_ENABLE | (orig & MSR_HV_SIEFP_RSVD_MASK);
782 WRMSR(MSR_HV_SIEFP, val);
783
784 /*
785 * Configure and unmask SINT for message and event flags.
786 */
787 sint = MSR_HV_SINT0 + VMBUS_SINT_MESSAGE;
788 orig = RDMSR(sint);
789 val = sc->vmbus_idtvec | MSR_HV_SINT_AUTOEOI |
790 (orig & MSR_HV_SINT_RSVD_MASK);
791 WRMSR(sint, val);
792
793 /*
794 * Configure and unmask SINT for timer.
795 */
796 vmbus_synic_setup1(sc);
797 /*
798 * All done; enable SynIC.
799 */
800 orig = RDMSR(MSR_HV_SCONTROL);
801 val = MSR_HV_SCTRL_ENABLE | (orig & MSR_HV_SCTRL_RSVD_MASK);
802 WRMSR(MSR_HV_SCONTROL, val);
803 }
804
805 #if defined(__x86_64__)
806 void
hyperv_vm_tlb_flush(pmap_t pmap,vm_offset_t addr1,vm_offset_t addr2,smp_invl_local_cb_t curcpu_cb,enum invl_op_codes op)807 hyperv_vm_tlb_flush(pmap_t pmap, vm_offset_t addr1, vm_offset_t addr2,
808 smp_invl_local_cb_t curcpu_cb, enum invl_op_codes op)
809 {
810 struct vmbus_softc *sc = vmbus_get_softc();
811 return hv_vm_tlb_flush(pmap, addr1, addr2, op, sc, curcpu_cb);
812 }
813 #endif /*__x86_64__*/
814
815 static void
vmbus_synic_teardown(void * arg)816 vmbus_synic_teardown(void *arg)
817 {
818 uint64_t orig;
819 uint32_t sint;
820
821 /*
822 * Disable SynIC.
823 */
824 orig = RDMSR(MSR_HV_SCONTROL);
825 WRMSR(MSR_HV_SCONTROL, (orig & MSR_HV_SCTRL_RSVD_MASK));
826
827 /*
828 * Mask message and event flags SINT.
829 */
830 sint = MSR_HV_SINT0 + VMBUS_SINT_MESSAGE;
831 orig = RDMSR(sint);
832 WRMSR(sint, orig | MSR_HV_SINT_MASKED);
833
834 /*
835 * Mask timer SINT.
836 */
837 vmbus_synic_teardown1();
838 /*
839 * Teardown SynIC message.
840 */
841 orig = RDMSR(MSR_HV_SIMP);
842 WRMSR(MSR_HV_SIMP, (orig & MSR_HV_SIMP_RSVD_MASK));
843
844 /*
845 * Teardown SynIC event flags.
846 */
847 orig = RDMSR(MSR_HV_SIEFP);
848 WRMSR(MSR_HV_SIEFP, (orig & MSR_HV_SIEFP_RSVD_MASK));
849 }
850
851 static int
vmbus_dma_alloc(struct vmbus_softc * sc)852 vmbus_dma_alloc(struct vmbus_softc *sc)
853 {
854 uint8_t *evtflags;
855 int cpu;
856
857 CPU_FOREACH(cpu) {
858 void *ptr;
859
860 /*
861 * Per-cpu messages and event flags.
862 */
863 ptr = contigmalloc(PAGE_SIZE, M_DEVBUF, M_WAITOK | M_ZERO,
864 0ul, ~0ul, PAGE_SIZE, 0);
865 if (ptr == NULL)
866 return ENOMEM;
867 VMBUS_PCPU_GET(sc, message, cpu) = ptr;
868
869 ptr = contigmalloc(PAGE_SIZE, M_DEVBUF, M_WAITOK | M_ZERO,
870 0ul, ~0ul, PAGE_SIZE, 0);
871 if (ptr == NULL)
872 return ENOMEM;
873 VMBUS_PCPU_GET(sc, event_flags, cpu) = ptr;
874 }
875
876 evtflags = contigmalloc(PAGE_SIZE, M_DEVBUF, M_WAITOK | M_ZERO,
877 0ul, ~0ul, PAGE_SIZE, 0);
878 if (evtflags == NULL)
879 return ENOMEM;
880 sc->vmbus_rx_evtflags = (u_long *)evtflags;
881 sc->vmbus_tx_evtflags = (u_long *)(evtflags + (PAGE_SIZE / 2));
882 sc->vmbus_evtflags = evtflags;
883
884 sc->vmbus_mnf1 = contigmalloc(PAGE_SIZE, M_DEVBUF, M_WAITOK | M_ZERO,
885 0ul, ~0ul, PAGE_SIZE, 0);
886 if (sc->vmbus_mnf1 == NULL)
887 return ENOMEM;
888
889 sc->vmbus_mnf2 = contigmalloc(sizeof(struct vmbus_mnf), M_DEVBUF,
890 M_WAITOK | M_ZERO, 0ul, ~0ul, PAGE_SIZE, 0);
891 if (sc->vmbus_mnf2 == NULL)
892 return ENOMEM;
893
894 return 0;
895 }
896
897 static void
vmbus_dma_free(struct vmbus_softc * sc)898 vmbus_dma_free(struct vmbus_softc *sc)
899 {
900 int cpu;
901
902 if (sc->vmbus_evtflags != NULL) {
903 free(sc->vmbus_evtflags, M_DEVBUF);
904 sc->vmbus_evtflags = NULL;
905 sc->vmbus_rx_evtflags = NULL;
906 sc->vmbus_tx_evtflags = NULL;
907 }
908 if (sc->vmbus_mnf1 != NULL) {
909 free(sc->vmbus_mnf1, M_DEVBUF);
910 sc->vmbus_mnf1 = NULL;
911 }
912 if (sc->vmbus_mnf2 != NULL) {
913 free(sc->vmbus_mnf2, M_DEVBUF);
914 sc->vmbus_mnf2 = NULL;
915 }
916
917 CPU_FOREACH(cpu) {
918 if (VMBUS_PCPU_GET(sc, message, cpu) != NULL) {
919 free(VMBUS_PCPU_GET(sc, message, cpu), M_DEVBUF);
920 VMBUS_PCPU_GET(sc, message, cpu) = NULL;
921 }
922 if (VMBUS_PCPU_GET(sc, event_flags, cpu) != NULL) {
923 free(VMBUS_PCPU_GET(sc, event_flags, cpu), M_DEVBUF);
924 VMBUS_PCPU_GET(sc, event_flags, cpu) = NULL;
925 }
926 }
927 }
928
929 static int
vmbus_intr_setup(struct vmbus_softc * sc)930 vmbus_intr_setup(struct vmbus_softc *sc)
931 {
932 int cpu;
933
934 CPU_FOREACH(cpu) {
935 char buf[MAXCOMLEN + 1];
936 cpuset_t cpu_mask;
937
938 /* Allocate an interrupt counter for Hyper-V interrupt */
939 snprintf(buf, sizeof(buf), "cpu%d:hyperv", cpu);
940 #if !defined(__aarch64__)
941 intrcnt_add(buf, VMBUS_PCPU_PTR(sc, intr_cnt, cpu));
942 #endif /* not for aarch64 */
943 /*
944 * Setup taskqueue to handle events. Task will be per-
945 * channel.
946 */
947 VMBUS_PCPU_GET(sc, event_tq, cpu) = taskqueue_create_fast(
948 "hyperv event", M_WAITOK, taskqueue_thread_enqueue,
949 VMBUS_PCPU_PTR(sc, event_tq, cpu));
950 if (vmbus_pin_evttask) {
951 CPU_SETOF(cpu, &cpu_mask);
952 taskqueue_start_threads_cpuset(
953 VMBUS_PCPU_PTR(sc, event_tq, cpu), 1, PI_NET,
954 &cpu_mask, "hvevent%d", cpu);
955 } else {
956 taskqueue_start_threads(
957 VMBUS_PCPU_PTR(sc, event_tq, cpu), 1, PI_NET,
958 "hvevent%d", cpu);
959 }
960
961 /*
962 * Setup tasks and taskqueues to handle messages.
963 */
964 VMBUS_PCPU_GET(sc, message_tq, cpu) = taskqueue_create_fast(
965 "hyperv msg", M_WAITOK, taskqueue_thread_enqueue,
966 VMBUS_PCPU_PTR(sc, message_tq, cpu));
967 CPU_SETOF(cpu, &cpu_mask);
968 taskqueue_start_threads_cpuset(
969 VMBUS_PCPU_PTR(sc, message_tq, cpu), 1, PI_NET, &cpu_mask,
970 "hvmsg%d", cpu);
971 TASK_INIT(VMBUS_PCPU_PTR(sc, message_task, cpu), 0,
972 vmbus_msg_task, sc);
973 }
974 return (vmbus_setup_intr1(sc));
975 }
976 static void
vmbus_intr_teardown(struct vmbus_softc * sc)977 vmbus_intr_teardown(struct vmbus_softc *sc)
978 {
979 vmbus_intr_teardown1(sc);
980 }
981
982 static int
vmbus_read_ivar(device_t dev,device_t child,int index,uintptr_t * result)983 vmbus_read_ivar(device_t dev, device_t child, int index, uintptr_t *result)
984 {
985 return (ENOENT);
986 }
987
988 static int
vmbus_child_pnpinfo(device_t dev,device_t child,struct sbuf * sb)989 vmbus_child_pnpinfo(device_t dev, device_t child, struct sbuf *sb)
990 {
991 const struct vmbus_channel *chan;
992 char guidbuf[HYPERV_GUID_STRLEN];
993
994 chan = vmbus_get_channel(child);
995 if (chan == NULL) {
996 /* Event timer device, which does not belong to a channel */
997 return (0);
998 }
999
1000 hyperv_guid2str(&chan->ch_guid_type, guidbuf, sizeof(guidbuf));
1001 sbuf_printf(sb, "classid=%s", guidbuf);
1002
1003 hyperv_guid2str(&chan->ch_guid_inst, guidbuf, sizeof(guidbuf));
1004 sbuf_printf(sb, " deviceid=%s", guidbuf);
1005
1006 return (0);
1007 }
1008
1009 int
vmbus_add_child(struct vmbus_channel * chan)1010 vmbus_add_child(struct vmbus_channel *chan)
1011 {
1012 struct vmbus_softc *sc = chan->ch_vmbus;
1013 device_t parent = sc->vmbus_dev;
1014
1015 bus_topo_lock();
1016
1017 chan->ch_dev = device_add_child(parent, NULL, DEVICE_UNIT_ANY);
1018 if (chan->ch_dev == NULL) {
1019 bus_topo_unlock();
1020 device_printf(parent, "device_add_child for chan%u failed\n",
1021 chan->ch_id);
1022 return (ENXIO);
1023 }
1024 device_set_ivars(chan->ch_dev, chan);
1025 device_probe_and_attach(chan->ch_dev);
1026 bus_topo_unlock();
1027
1028 return (0);
1029 }
1030
1031 int
vmbus_delete_child(struct vmbus_channel * chan)1032 vmbus_delete_child(struct vmbus_channel *chan)
1033 {
1034 int error = 0;
1035
1036 bus_topo_lock();
1037 if (chan->ch_dev != NULL) {
1038 error = device_delete_child(chan->ch_vmbus->vmbus_dev,
1039 chan->ch_dev);
1040 chan->ch_dev = NULL;
1041 }
1042 bus_topo_unlock();
1043 return (error);
1044 }
1045
1046 static int
vmbus_sysctl_version(SYSCTL_HANDLER_ARGS)1047 vmbus_sysctl_version(SYSCTL_HANDLER_ARGS)
1048 {
1049 struct vmbus_softc *sc = arg1;
1050 char verstr[16];
1051
1052 snprintf(verstr, sizeof(verstr), "%u.%u",
1053 VMBUS_VERSION_MAJOR(sc->vmbus_version),
1054 VMBUS_VERSION_MINOR(sc->vmbus_version));
1055 return sysctl_handle_string(oidp, verstr, sizeof(verstr), req);
1056 }
1057
1058 /*
1059 * We need the function to make sure the MMIO resource is allocated from the
1060 * ranges found in _CRS.
1061 *
1062 * For the release function, we can use bus_generic_release_resource().
1063 */
1064 static struct resource *
vmbus_alloc_resource(device_t dev,device_t child,int type,int * rid,rman_res_t start,rman_res_t end,rman_res_t count,u_int flags)1065 vmbus_alloc_resource(device_t dev, device_t child, int type, int *rid,
1066 rman_res_t start, rman_res_t end, rman_res_t count, u_int flags)
1067 {
1068 device_t parent = device_get_parent(dev);
1069 struct resource *res;
1070
1071 if (type == SYS_RES_MEMORY) {
1072 struct vmbus_softc *sc = device_get_softc(dev);
1073
1074 res = pcib_host_res_alloc(&sc->vmbus_mmio_res, child, type,
1075 rid, start, end, count, flags);
1076 } else {
1077 res = BUS_ALLOC_RESOURCE(parent, child, type, rid, start,
1078 end, count, flags);
1079 }
1080
1081 return (res);
1082 }
1083
1084 static int
vmbus_alloc_msi(device_t bus,device_t dev,int count,int maxcount,int * irqs)1085 vmbus_alloc_msi(device_t bus, device_t dev, int count, int maxcount, int *irqs)
1086 {
1087
1088 return (PCIB_ALLOC_MSI(device_get_parent(bus), dev, count, maxcount,
1089 irqs));
1090 }
1091
1092 static int
vmbus_release_msi(device_t bus,device_t dev,int count,int * irqs)1093 vmbus_release_msi(device_t bus, device_t dev, int count, int *irqs)
1094 {
1095
1096 return (PCIB_RELEASE_MSI(device_get_parent(bus), dev, count, irqs));
1097 }
1098
1099 static int
vmbus_alloc_msix(device_t bus,device_t dev,int * irq)1100 vmbus_alloc_msix(device_t bus, device_t dev, int *irq)
1101 {
1102
1103 return (PCIB_ALLOC_MSIX(device_get_parent(bus), dev, irq));
1104 }
1105
1106 static int
vmbus_release_msix(device_t bus,device_t dev,int irq)1107 vmbus_release_msix(device_t bus, device_t dev, int irq)
1108 {
1109
1110 return (PCIB_RELEASE_MSIX(device_get_parent(bus), dev, irq));
1111 }
1112
1113 static int
vmbus_map_msi(device_t bus,device_t dev,int irq,uint64_t * addr,uint32_t * data)1114 vmbus_map_msi(device_t bus, device_t dev, int irq, uint64_t *addr,
1115 uint32_t *data)
1116 {
1117
1118 return (PCIB_MAP_MSI(device_get_parent(bus), dev, irq, addr, data));
1119 }
1120
1121 static uint32_t
vmbus_get_version_method(device_t bus,device_t dev)1122 vmbus_get_version_method(device_t bus, device_t dev)
1123 {
1124 struct vmbus_softc *sc = device_get_softc(bus);
1125
1126 return sc->vmbus_version;
1127 }
1128
1129 static int
vmbus_probe_guid_method(device_t bus,device_t dev,const struct hyperv_guid * guid)1130 vmbus_probe_guid_method(device_t bus, device_t dev,
1131 const struct hyperv_guid *guid)
1132 {
1133 const struct vmbus_channel *chan = vmbus_get_channel(dev);
1134
1135 if (memcmp(&chan->ch_guid_type, guid, sizeof(struct hyperv_guid)) == 0)
1136 return 0;
1137 return ENXIO;
1138 }
1139
1140 static uint32_t
vmbus_get_vcpu_id_method(device_t bus,device_t dev,int cpu)1141 vmbus_get_vcpu_id_method(device_t bus, device_t dev, int cpu)
1142 {
1143 const struct vmbus_softc *sc = device_get_softc(bus);
1144
1145 return (VMBUS_PCPU_GET(sc, vcpuid, cpu));
1146 }
1147
1148 static struct taskqueue *
vmbus_get_eventtq_method(device_t bus,device_t dev __unused,int cpu)1149 vmbus_get_eventtq_method(device_t bus, device_t dev __unused, int cpu)
1150 {
1151 const struct vmbus_softc *sc = device_get_softc(bus);
1152
1153 KASSERT(cpu >= 0 && cpu < mp_ncpus, ("invalid cpu%d", cpu));
1154 return (VMBUS_PCPU_GET(sc, event_tq, cpu));
1155 }
1156
1157 #define VTPM_BASE_ADDR 0xfed40000
1158 #define FOUR_GB (1ULL << 32)
1159
1160 enum parse_pass { parse_64, parse_32 };
1161
1162 struct parse_context {
1163 device_t vmbus_dev;
1164 enum parse_pass pass;
1165 };
1166
1167 static ACPI_STATUS
parse_crs(ACPI_RESOURCE * res,void * ctx)1168 parse_crs(ACPI_RESOURCE *res, void *ctx)
1169 {
1170 const struct parse_context *pc = ctx;
1171 device_t vmbus_dev = pc->vmbus_dev;
1172
1173 struct vmbus_softc *sc = device_get_softc(vmbus_dev);
1174 UINT64 start, end;
1175
1176 switch (res->Type) {
1177 case ACPI_RESOURCE_TYPE_ADDRESS32:
1178 start = res->Data.Address32.Address.Minimum;
1179 end = res->Data.Address32.Address.Maximum;
1180 break;
1181
1182 case ACPI_RESOURCE_TYPE_ADDRESS64:
1183 start = res->Data.Address64.Address.Minimum;
1184 end = res->Data.Address64.Address.Maximum;
1185 break;
1186
1187 default:
1188 /* Unused types. */
1189 return (AE_OK);
1190 }
1191
1192 /*
1193 * We don't use <1MB addresses.
1194 */
1195 if (end < 0x100000)
1196 return (AE_OK);
1197
1198 /* Don't conflict with vTPM. */
1199 if (end >= VTPM_BASE_ADDR && start < VTPM_BASE_ADDR)
1200 end = VTPM_BASE_ADDR - 1;
1201
1202 if ((pc->pass == parse_32 && start < FOUR_GB) ||
1203 (pc->pass == parse_64 && start >= FOUR_GB))
1204 pcib_host_res_decodes(&sc->vmbus_mmio_res, SYS_RES_MEMORY,
1205 start, end, 0);
1206
1207 return (AE_OK);
1208 }
1209
1210 static void
vmbus_get_crs(device_t dev,device_t vmbus_dev,enum parse_pass pass)1211 vmbus_get_crs(device_t dev, device_t vmbus_dev, enum parse_pass pass)
1212 {
1213 struct parse_context pc;
1214 ACPI_STATUS status;
1215
1216 if (bootverbose)
1217 device_printf(dev, "walking _CRS, pass=%d\n", pass);
1218
1219 pc.vmbus_dev = vmbus_dev;
1220 pc.pass = pass;
1221 status = AcpiWalkResources(acpi_get_handle(dev), "_CRS",
1222 parse_crs, &pc);
1223
1224 if (bootverbose && ACPI_FAILURE(status))
1225 device_printf(dev, "_CRS: not found, pass=%d\n", pass);
1226 }
1227
1228 static void
vmbus_get_mmio_res_pass(device_t dev,enum parse_pass pass)1229 vmbus_get_mmio_res_pass(device_t dev, enum parse_pass pass)
1230 {
1231 device_t acpi0, parent;
1232
1233 parent = device_get_parent(dev);
1234
1235 acpi0 = device_get_parent(parent);
1236 if (strcmp("acpi0", device_get_nameunit(acpi0)) == 0) {
1237 device_t *children;
1238 int count;
1239
1240 /*
1241 * Try to locate VMBUS resources and find _CRS on them.
1242 */
1243 if (device_get_children(acpi0, &children, &count) == 0) {
1244 int i;
1245
1246 for (i = 0; i < count; ++i) {
1247 if (!device_is_attached(children[i]))
1248 continue;
1249
1250 if (strcmp("vmbus_res",
1251 device_get_name(children[i])) == 0)
1252 vmbus_get_crs(children[i], dev, pass);
1253 }
1254 free(children, M_TEMP);
1255 }
1256
1257 /*
1258 * Try to find _CRS on acpi.
1259 */
1260 vmbus_get_crs(acpi0, dev, pass);
1261 } else {
1262 device_printf(dev, "not grandchild of acpi\n");
1263 }
1264
1265 /*
1266 * Try to find _CRS on parent.
1267 */
1268 vmbus_get_crs(parent, dev, pass);
1269 }
1270
1271 static void
vmbus_get_mmio_res(device_t dev)1272 vmbus_get_mmio_res(device_t dev)
1273 {
1274 struct vmbus_softc *sc = device_get_softc(dev);
1275 /*
1276 * We walk the resources twice to make sure that: in the resource
1277 * list, the 32-bit resources appear behind the 64-bit resources.
1278 * NB: resource_list_add() uses INSERT_TAIL. This way, when we
1279 * iterate through the list to find a range for a 64-bit BAR in
1280 * vmbus_alloc_resource(), we can make sure we try to use >4GB
1281 * ranges first.
1282 */
1283 pcib_host_res_init(dev, &sc->vmbus_mmio_res);
1284
1285 vmbus_get_mmio_res_pass(dev, parse_64);
1286 vmbus_get_mmio_res_pass(dev, parse_32);
1287 }
1288
1289 /*
1290 * On Gen2 VMs, Hyper-V provides mmio space for framebuffer.
1291 * This mmio address range is not useable for other PCI devices.
1292 * Currently only efifb and vbefb drivers are using this range without
1293 * reserving it from system.
1294 * Therefore, vmbus driver reserves it before any other PCI device
1295 * drivers start to request mmio addresses.
1296 */
1297 static struct resource *hv_fb_res;
1298
1299 static void
vmbus_fb_mmio_res(device_t dev)1300 vmbus_fb_mmio_res(device_t dev)
1301 {
1302 struct efi_fb *efifb;
1303 #if !defined(__aarch64__)
1304 struct vbe_fb *vbefb;
1305 #endif /* aarch64 */
1306 rman_res_t fb_start, fb_end, fb_count;
1307 int fb_height, fb_width;
1308 caddr_t kmdp;
1309
1310 struct vmbus_softc *sc = device_get_softc(dev);
1311 int rid = 0;
1312
1313 kmdp = preload_search_by_type("elf kernel");
1314 if (kmdp == NULL)
1315 kmdp = preload_search_by_type("elf64 kernel");
1316 efifb = (struct efi_fb *)preload_search_info(kmdp,
1317 MODINFO_METADATA | MODINFOMD_EFI_FB);
1318 #if !defined(__aarch64__)
1319 vbefb = (struct vbe_fb *)preload_search_info(kmdp,
1320 MODINFO_METADATA | MODINFOMD_VBE_FB);
1321 #endif /* aarch64 */
1322 if (efifb != NULL) {
1323 fb_start = efifb->fb_addr;
1324 fb_end = efifb->fb_addr + efifb->fb_size;
1325 fb_count = efifb->fb_size;
1326 fb_height = efifb->fb_height;
1327 fb_width = efifb->fb_width;
1328 }
1329 #if !defined(__aarch64__)
1330 else if (vbefb != NULL) {
1331 fb_start = vbefb->fb_addr;
1332 fb_end = vbefb->fb_addr + vbefb->fb_size;
1333 fb_count = vbefb->fb_size;
1334 fb_height = vbefb->fb_height;
1335 fb_width = vbefb->fb_width;
1336 }
1337 #endif /* aarch64 */
1338 else {
1339 if (bootverbose)
1340 device_printf(dev,
1341 "no preloaded kernel fb information\n");
1342 /* We are on Gen1 VM, just return. */
1343 return;
1344 }
1345
1346 if (bootverbose)
1347 device_printf(dev,
1348 "fb: fb_addr: %#jx, size: %#jx, "
1349 "actual size needed: 0x%x\n",
1350 fb_start, fb_count, fb_height * fb_width);
1351
1352 hv_fb_res = pcib_host_res_alloc(&sc->vmbus_mmio_res, dev,
1353 SYS_RES_MEMORY, &rid, fb_start, fb_end, fb_count,
1354 RF_ACTIVE | rman_make_alignment_flags(PAGE_SIZE));
1355
1356 if (hv_fb_res && bootverbose)
1357 device_printf(dev,
1358 "successfully reserved memory for framebuffer "
1359 "starting at %#jx, size %#jx\n",
1360 fb_start, fb_count);
1361 }
1362
1363 static void
vmbus_free_mmio_res(device_t dev)1364 vmbus_free_mmio_res(device_t dev)
1365 {
1366 struct vmbus_softc *sc = device_get_softc(dev);
1367
1368 pcib_host_res_free(dev, &sc->vmbus_mmio_res);
1369
1370 if (hv_fb_res)
1371 hv_fb_res = NULL;
1372 }
1373
1374 static void
vmbus_identify(driver_t * driver,device_t parent)1375 vmbus_identify(driver_t *driver, device_t parent)
1376 {
1377
1378 if (device_get_unit(parent) != 0 || vm_guest != VM_GUEST_HV ||
1379 (hyperv_features & CPUID_HV_MSR_SYNIC) == 0)
1380 return;
1381 device_add_child(parent, "vmbus", DEVICE_UNIT_ANY);
1382 }
1383
1384 static int
vmbus_probe(device_t dev)1385 vmbus_probe(device_t dev)
1386 {
1387
1388 if (device_get_unit(dev) != 0 || vm_guest != VM_GUEST_HV ||
1389 (hyperv_features & CPUID_HV_MSR_SYNIC) == 0)
1390 return (ENXIO);
1391
1392 device_set_desc(dev, "Hyper-V Vmbus");
1393 return (BUS_PROBE_DEFAULT);
1394 }
1395
1396 #if defined(__x86_64__)
1397 static int
vmbus_alloc_cpu_mem(struct vmbus_softc * sc)1398 vmbus_alloc_cpu_mem(struct vmbus_softc *sc)
1399 {
1400 int cpu;
1401
1402 CPU_FOREACH(cpu) {
1403 void **hv_cpu_mem;
1404
1405 hv_cpu_mem = VMBUS_PCPU_PTR(sc, cpu_mem, cpu);
1406 *hv_cpu_mem = contigmalloc(PAGE_SIZE, M_DEVBUF,
1407 M_NOWAIT | M_ZERO, 0ul, ~0ul, PAGE_SIZE, 0);
1408
1409 if (*hv_cpu_mem == NULL)
1410 return ENOMEM;
1411 }
1412
1413 return 0;
1414 }
1415
1416 static void
vmbus_free_cpu_mem(struct vmbus_softc * sc)1417 vmbus_free_cpu_mem(struct vmbus_softc *sc)
1418 {
1419 int cpu;
1420
1421 CPU_FOREACH(cpu) {
1422 void **hv_cpu_mem;
1423 hv_cpu_mem = VMBUS_PCPU_PTR(sc, cpu_mem, cpu);
1424 if(*hv_cpu_mem != NULL) {
1425 free(*hv_cpu_mem, M_DEVBUF);
1426 *hv_cpu_mem = NULL;
1427 }
1428 }
1429 }
1430 #endif
1431
1432 /**
1433 * @brief Main vmbus driver initialization routine.
1434 *
1435 * Here, we
1436 * - initialize the vmbus driver context
1437 * - setup various driver entry points
1438 * - invoke the vmbus hv main init routine
1439 * - get the irq resource
1440 * - invoke the vmbus to add the vmbus root device
1441 * - setup the vmbus root device
1442 * - retrieve the channel offers
1443 */
1444 static int
vmbus_doattach(struct vmbus_softc * sc)1445 vmbus_doattach(struct vmbus_softc *sc)
1446 {
1447 struct sysctl_oid_list *child;
1448 struct sysctl_ctx_list *ctx;
1449 int ret;
1450 device_t dev_res;
1451 ACPI_HANDLE handle;
1452 unsigned int coherent = 0;
1453
1454 if (sc->vmbus_flags & VMBUS_FLAG_ATTACHED)
1455 return (0);
1456
1457 vmbus_get_mmio_res(sc->vmbus_dev);
1458 vmbus_fb_mmio_res(sc->vmbus_dev);
1459
1460 sc->vmbus_flags |= VMBUS_FLAG_ATTACHED;
1461
1462 sc->vmbus_gpadl = VMBUS_GPADL_START;
1463 mtx_init(&sc->vmbus_prichan_lock, "vmbus prichan", NULL, MTX_DEF);
1464 TAILQ_INIT(&sc->vmbus_prichans);
1465 mtx_init(&sc->vmbus_chan_lock, "vmbus channel", NULL, MTX_DEF);
1466 TAILQ_INIT(&sc->vmbus_chans);
1467 sc->vmbus_chmap = malloc(
1468 sizeof(struct vmbus_channel *) * VMBUS_CHAN_MAX, M_DEVBUF,
1469 M_WAITOK | M_ZERO);
1470
1471 /* Coherency attribute */
1472 dev_res = devclass_get_device(devclass_find("vmbus_res"), 0);
1473 if (dev_res != NULL) {
1474 handle = acpi_get_handle(dev_res);
1475
1476 if (ACPI_FAILURE(acpi_GetInteger(handle, "_CCA", &coherent)))
1477 coherent = 0;
1478 }
1479 if (bootverbose)
1480 device_printf(sc->vmbus_dev, "Bus is%s cache-coherent\n",
1481 coherent ? "" : " not");
1482
1483 bus_dma_tag_create(bus_get_dma_tag(sc->vmbus_dev),
1484 1, 0,
1485 BUS_SPACE_MAXADDR,
1486 BUS_SPACE_MAXADDR,
1487 NULL, NULL,
1488 BUS_SPACE_MAXSIZE,
1489 BUS_SPACE_UNRESTRICTED,
1490 BUS_SPACE_MAXSIZE,
1491 coherent ? BUS_DMA_COHERENT : 0,
1492 NULL, NULL,
1493 &sc->dmat);
1494 /*
1495 * Create context for "post message" Hypercalls
1496 */
1497 sc->vmbus_xc = vmbus_xact_ctx_create(bus_get_dma_tag(sc->vmbus_dev),
1498 HYPERCALL_POSTMSGIN_SIZE, VMBUS_MSG_SIZE,
1499 sizeof(struct vmbus_msghc));
1500 if (sc->vmbus_xc == NULL) {
1501 ret = ENXIO;
1502 goto cleanup;
1503 }
1504
1505 /*
1506 * Allocate DMA stuffs.
1507 */
1508 ret = vmbus_dma_alloc(sc);
1509 if (ret != 0)
1510 goto cleanup;
1511
1512 /*
1513 * Setup interrupt.
1514 */
1515 ret = vmbus_intr_setup(sc);
1516 if (ret != 0)
1517 goto cleanup;
1518
1519 #if defined(__x86_64__)
1520 /*
1521 * Alloc per cpu memory for tlb flush hypercall
1522 */
1523 if (hv_tlb_hcall) {
1524 ret = vmbus_alloc_cpu_mem(sc);
1525 if (ret != 0) {
1526 hv_tlb_hcall = 0;
1527 if (bootverbose)
1528 device_printf(sc->vmbus_dev,
1529 "cannot alloc contig memory for "
1530 "cpu_mem, use system provided "
1531 "tlb flush call.\n");
1532
1533 vmbus_free_cpu_mem(sc);
1534 }
1535 }
1536 #endif
1537
1538 /*
1539 * Setup SynIC.
1540 */
1541 if (bootverbose)
1542 device_printf(sc->vmbus_dev, "smp_started = %d\n", smp_started);
1543 smp_rendezvous(NULL, vmbus_synic_setup, NULL, sc);
1544 sc->vmbus_flags |= VMBUS_FLAG_SYNIC;
1545
1546 #if defined(__x86_64__)
1547 if (hv_tlb_hcall)
1548 smp_targeted_tlb_shootdown = &hyperv_vm_tlb_flush;
1549 #endif
1550
1551 /*
1552 * Initialize vmbus, e.g. connect to Hypervisor.
1553 */
1554 ret = vmbus_init(sc);
1555 if (ret != 0)
1556 goto cleanup;
1557
1558 if (sc->vmbus_version == VMBUS_VERSION_WS2008 ||
1559 sc->vmbus_version == VMBUS_VERSION_WIN7)
1560 sc->vmbus_event_proc = vmbus_event_proc_compat;
1561 else
1562 sc->vmbus_event_proc = vmbus_event_proc;
1563
1564 ret = vmbus_scan(sc);
1565 if (ret != 0)
1566 goto cleanup;
1567
1568 ctx = device_get_sysctl_ctx(sc->vmbus_dev);
1569 child = SYSCTL_CHILDREN(device_get_sysctl_tree(sc->vmbus_dev));
1570 SYSCTL_ADD_PROC(ctx, child, OID_AUTO, "version",
1571 CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, sc, 0,
1572 vmbus_sysctl_version, "A", "vmbus version");
1573
1574 return (ret);
1575
1576 cleanup:
1577 vmbus_scan_teardown(sc);
1578 vmbus_intr_teardown(sc);
1579 vmbus_dma_free(sc);
1580 if (sc->vmbus_xc != NULL) {
1581 vmbus_xact_ctx_destroy(sc->vmbus_xc);
1582 sc->vmbus_xc = NULL;
1583 }
1584 free(__DEVOLATILE(void *, sc->vmbus_chmap), M_DEVBUF);
1585 mtx_destroy(&sc->vmbus_prichan_lock);
1586 mtx_destroy(&sc->vmbus_chan_lock);
1587
1588 return (ret);
1589 }
1590
1591 static void
vmbus_event_proc_dummy(struct vmbus_softc * sc __unused,int cpu __unused)1592 vmbus_event_proc_dummy(struct vmbus_softc *sc __unused, int cpu __unused)
1593 {
1594 }
1595
1596 #if defined(EARLY_AP_STARTUP)
1597
1598 static void
vmbus_intrhook(void * xsc)1599 vmbus_intrhook(void *xsc)
1600 {
1601 struct vmbus_softc *sc = xsc;
1602
1603 if (bootverbose)
1604 device_printf(sc->vmbus_dev, "intrhook\n");
1605 vmbus_doattach(sc);
1606 config_intrhook_disestablish(&sc->vmbus_intrhook);
1607 }
1608
1609 #endif /* EARLY_AP_STARTUP */
1610
1611 static int
vmbus_attach(device_t dev)1612 vmbus_attach(device_t dev)
1613 {
1614 vmbus_sc = device_get_softc(dev);
1615 vmbus_sc->vmbus_dev = dev;
1616 vmbus_sc->vmbus_idtvec = -1;
1617
1618 /*
1619 * Event processing logic will be configured:
1620 * - After the vmbus protocol version negotiation.
1621 * - Before we request channel offers.
1622 */
1623 vmbus_sc->vmbus_event_proc = vmbus_event_proc_dummy;
1624
1625 #if defined(EARLY_AP_STARTUP)
1626 /*
1627 * Defer the real attach until the pause(9) works as expected.
1628 */
1629 vmbus_sc->vmbus_intrhook.ich_func = vmbus_intrhook;
1630 vmbus_sc->vmbus_intrhook.ich_arg = vmbus_sc;
1631 config_intrhook_establish(&vmbus_sc->vmbus_intrhook);
1632 #endif /* EARLY_AP_STARTUP and aarch64 */
1633
1634 return (0);
1635 }
1636
1637 static int
vmbus_detach(device_t dev)1638 vmbus_detach(device_t dev)
1639 {
1640 struct vmbus_softc *sc = device_get_softc(dev);
1641
1642 bus_generic_detach(dev);
1643 vmbus_chan_destroy_all(sc);
1644
1645 vmbus_scan_teardown(sc);
1646
1647 vmbus_disconnect(sc);
1648
1649 if (sc->vmbus_flags & VMBUS_FLAG_SYNIC) {
1650 sc->vmbus_flags &= ~VMBUS_FLAG_SYNIC;
1651 smp_rendezvous(NULL, vmbus_synic_teardown, NULL, NULL);
1652 }
1653
1654 #if defined(__x86_64__)
1655 /*
1656 * Restore the tlb flush to native call
1657 */
1658 if (hv_tlb_hcall) {
1659 smp_targeted_tlb_shootdown = &smp_targeted_tlb_shootdown_native;
1660 vmbus_free_cpu_mem(sc);
1661 }
1662 #endif
1663
1664 vmbus_intr_teardown(sc);
1665 vmbus_dma_free(sc);
1666
1667 if (sc->vmbus_xc != NULL) {
1668 vmbus_xact_ctx_destroy(sc->vmbus_xc);
1669 sc->vmbus_xc = NULL;
1670 }
1671
1672 free(__DEVOLATILE(void *, sc->vmbus_chmap), M_DEVBUF);
1673 mtx_destroy(&sc->vmbus_prichan_lock);
1674 mtx_destroy(&sc->vmbus_chan_lock);
1675
1676 vmbus_free_mmio_res(dev);
1677
1678 #if defined(__aarch64__)
1679 bus_release_resource(device_get_parent(dev), SYS_RES_IRQ, sc->vector,
1680 sc->ires);
1681 #endif
1682 return (0);
1683 }
1684
1685 #if !defined(EARLY_AP_STARTUP)
1686
1687 static void
vmbus_sysinit(void * arg __unused)1688 vmbus_sysinit(void *arg __unused)
1689 {
1690 struct vmbus_softc *sc = vmbus_get_softc();
1691
1692 if (vm_guest != VM_GUEST_HV || sc == NULL)
1693 return;
1694
1695 vmbus_doattach(sc);
1696 }
1697 /*
1698 * NOTE:
1699 * We have to start as the last step of SI_SUB_SMP, i.e. after SMP is
1700 * initialized.
1701 */
1702 SYSINIT(vmbus_initialize, SI_SUB_SMP, SI_ORDER_ANY, vmbus_sysinit, NULL);
1703 #endif /* !EARLY_AP_STARTUP */
1704