1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2010,2013 Lawrence Stewart <lstewart@freebsd.org> 5 * Copyright (c) 2010 The FreeBSD Foundation 6 * All rights reserved. 7 * 8 * This software was developed by Lawrence Stewart while studying at the Centre 9 * for Advanced Internet Architectures, Swinburne University of Technology, 10 * made possible in part by grants from the FreeBSD Foundation and Cisco 11 * University Research Program Fund at Community Foundation Silicon Valley. 12 * 13 * Portions of this software were developed at the Centre for Advanced 14 * Internet Architectures, Swinburne University of Technology, Melbourne, 15 * Australia by Lawrence Stewart under sponsorship from the FreeBSD Foundation. 16 * 17 * Redistribution and use in source and binary forms, with or without 18 * modification, are permitted provided that the following conditions 19 * are met: 20 * 1. Redistributions of source code must retain the above copyright 21 * notice, this list of conditions and the following disclaimer. 22 * 2. Redistributions in binary form must reproduce the above copyright 23 * notice, this list of conditions and the following disclaimer in the 24 * documentation and/or other materials provided with the distribution. 25 * 26 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 27 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 28 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 29 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 30 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 31 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 32 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 33 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 34 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 35 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 36 * SUCH DAMAGE. 37 */ 38 39 #include <sys/param.h> 40 #include <sys/kernel.h> 41 #include <sys/hhook.h> 42 #include <sys/khelp.h> 43 #include <sys/malloc.h> 44 #include <sys/module.h> 45 #include <sys/module_khelp.h> 46 #include <sys/osd.h> 47 #include <sys/queue.h> 48 #include <sys/refcount.h> 49 #include <sys/systm.h> 50 51 #include <net/vnet.h> 52 53 struct hhook { 54 hhook_func_t hhk_func; 55 struct helper *hhk_helper; 56 void *hhk_udata; 57 STAILQ_ENTRY(hhook) hhk_next; 58 }; 59 60 static MALLOC_DEFINE(M_HHOOK, "hhook", "Helper hooks are linked off hhook_head lists"); 61 62 LIST_HEAD(hhookheadhead, hhook_head); 63 struct hhookheadhead hhook_head_list; 64 VNET_DEFINE(struct hhookheadhead, hhook_vhead_list); 65 #define V_hhook_vhead_list VNET(hhook_vhead_list) 66 67 static struct mtx hhook_head_list_lock; 68 MTX_SYSINIT(hhookheadlistlock, &hhook_head_list_lock, "hhook_head list lock", 69 MTX_DEF); 70 71 /* Protected by hhook_head_list_lock. */ 72 static uint32_t n_hhookheads; 73 74 /* Private function prototypes. */ 75 static void hhook_head_destroy(struct hhook_head *hhh); 76 void khelp_new_hhook_registered(struct hhook_head *hhh, uint32_t flags); 77 78 #define HHHLIST_LOCK() mtx_lock(&hhook_head_list_lock) 79 #define HHHLIST_UNLOCK() mtx_unlock(&hhook_head_list_lock) 80 #define HHHLIST_LOCK_ASSERT() mtx_assert(&hhook_head_list_lock, MA_OWNED) 81 82 #define HHH_LOCK_INIT(hhh) rm_init(&(hhh)->hhh_lock, "hhook_head rm lock") 83 #define HHH_LOCK_DESTROY(hhh) rm_destroy(&(hhh)->hhh_lock) 84 #define HHH_WLOCK(hhh) rm_wlock(&(hhh)->hhh_lock) 85 #define HHH_WUNLOCK(hhh) rm_wunlock(&(hhh)->hhh_lock) 86 #define HHH_RLOCK(hhh, rmpt) rm_rlock(&(hhh)->hhh_lock, (rmpt)) 87 #define HHH_RUNLOCK(hhh, rmpt) rm_runlock(&(hhh)->hhh_lock, (rmpt)) 88 89 /* 90 * Run all helper hook functions for a given hook point. 91 */ 92 void 93 hhook_run_hooks(struct hhook_head *hhh, void *ctx_data, struct osd *hosd) 94 { 95 struct hhook *hhk; 96 void *hdata; 97 struct rm_priotracker rmpt; 98 99 KASSERT(hhh->hhh_refcount > 0, ("hhook_head %p refcount is 0", hhh)); 100 101 HHH_RLOCK(hhh, &rmpt); 102 STAILQ_FOREACH(hhk, &hhh->hhh_hooks, hhk_next) { 103 if (hhk->hhk_helper != NULL && 104 hhk->hhk_helper->h_flags & HELPER_NEEDS_OSD) { 105 hdata = osd_get(OSD_KHELP, hosd, hhk->hhk_helper->h_id); 106 if (hdata == NULL) 107 continue; 108 } else 109 hdata = NULL; 110 111 /* 112 * XXXLAS: We currently ignore the int returned by the hook, 113 * but will likely want to handle it in future to allow hhook to 114 * be used like pfil and effect changes at the hhook calling 115 * site e.g. we could define a new hook type of HHOOK_TYPE_PFIL 116 * and standardise what particular return values mean and set 117 * the context data to pass exactly the same information as pfil 118 * hooks currently receive, thus replicating pfil with hhook. 119 */ 120 hhk->hhk_func(hhh->hhh_type, hhh->hhh_id, hhk->hhk_udata, 121 ctx_data, hdata, hosd); 122 } 123 HHH_RUNLOCK(hhh, &rmpt); 124 } 125 126 /* 127 * Register a new helper hook function with a helper hook point. 128 */ 129 int 130 hhook_add_hook(struct hhook_head *hhh, struct hookinfo *hki, uint32_t flags) 131 { 132 struct hhook *hhk, *tmp; 133 int error; 134 135 error = 0; 136 137 if (hhh == NULL) 138 return (ENOENT); 139 140 hhk = malloc(sizeof(struct hhook), M_HHOOK, 141 M_ZERO | ((flags & HHOOK_WAITOK) ? M_WAITOK : M_NOWAIT)); 142 143 if (hhk == NULL) 144 return (ENOMEM); 145 146 hhk->hhk_helper = hki->hook_helper; 147 hhk->hhk_func = hki->hook_func; 148 hhk->hhk_udata = hki->hook_udata; 149 150 HHH_WLOCK(hhh); 151 STAILQ_FOREACH(tmp, &hhh->hhh_hooks, hhk_next) { 152 if (tmp->hhk_func == hki->hook_func && 153 tmp->hhk_udata == hki->hook_udata) { 154 /* The helper hook function is already registered. */ 155 error = EEXIST; 156 break; 157 } 158 } 159 160 if (!error) { 161 STAILQ_INSERT_TAIL(&hhh->hhh_hooks, hhk, hhk_next); 162 hhh->hhh_nhooks++; 163 } else 164 free(hhk, M_HHOOK); 165 166 HHH_WUNLOCK(hhh); 167 168 return (error); 169 } 170 171 /* 172 * Register a helper hook function with a helper hook point (including all 173 * virtual instances of the hook point if it is virtualised). 174 * 175 * The logic is unfortunately far more complex than for 176 * hhook_remove_hook_lookup() because hhook_add_hook() can call malloc() with 177 * M_WAITOK and thus we cannot call hhook_add_hook() with the 178 * hhook_head_list_lock held. 179 * 180 * The logic assembles an array of hhook_head structs that correspond to the 181 * helper hook point being hooked and bumps the refcount on each (all done with 182 * the hhook_head_list_lock held). The hhook_head_list_lock is then dropped, and 183 * hhook_add_hook() is called and the refcount dropped for each hhook_head 184 * struct in the array. 185 */ 186 int 187 hhook_add_hook_lookup(struct hookinfo *hki, uint32_t flags) 188 { 189 struct hhook_head **heads_to_hook, *hhh; 190 int error, i, n_heads_to_hook; 191 192 tryagain: 193 error = i = 0; 194 /* 195 * Accessing n_hhookheads without hhook_head_list_lock held opens up a 196 * race with hhook_head_register() which we are unlikely to lose, but 197 * nonetheless have to cope with - hence the complex goto logic. 198 */ 199 n_heads_to_hook = n_hhookheads; 200 heads_to_hook = malloc(n_heads_to_hook * sizeof(struct hhook_head *), 201 M_HHOOK, flags & HHOOK_WAITOK ? M_WAITOK : M_NOWAIT); 202 if (heads_to_hook == NULL) 203 return (ENOMEM); 204 205 HHHLIST_LOCK(); 206 LIST_FOREACH(hhh, &hhook_head_list, hhh_next) { 207 if (hhh->hhh_type == hki->hook_type && 208 hhh->hhh_id == hki->hook_id) { 209 if (i < n_heads_to_hook) { 210 heads_to_hook[i] = hhh; 211 refcount_acquire(&heads_to_hook[i]->hhh_refcount); 212 i++; 213 } else { 214 /* 215 * We raced with hhook_head_register() which 216 * inserted a hhook_head that we need to hook 217 * but did not malloc space for. Abort this run 218 * and try again. 219 */ 220 for (i--; i >= 0; i--) 221 refcount_release(&heads_to_hook[i]->hhh_refcount); 222 free(heads_to_hook, M_HHOOK); 223 HHHLIST_UNLOCK(); 224 goto tryagain; 225 } 226 } 227 } 228 HHHLIST_UNLOCK(); 229 230 for (i--; i >= 0; i--) { 231 if (!error) 232 error = hhook_add_hook(heads_to_hook[i], hki, flags); 233 refcount_release(&heads_to_hook[i]->hhh_refcount); 234 } 235 236 free(heads_to_hook, M_HHOOK); 237 238 return (error); 239 } 240 241 /* 242 * Remove a helper hook function from a helper hook point. 243 */ 244 int 245 hhook_remove_hook(struct hhook_head *hhh, struct hookinfo *hki) 246 { 247 struct hhook *tmp; 248 249 if (hhh == NULL) 250 return (ENOENT); 251 252 HHH_WLOCK(hhh); 253 STAILQ_FOREACH(tmp, &hhh->hhh_hooks, hhk_next) { 254 if (tmp->hhk_func == hki->hook_func && 255 tmp->hhk_udata == hki->hook_udata) { 256 STAILQ_REMOVE(&hhh->hhh_hooks, tmp, hhook, hhk_next); 257 free(tmp, M_HHOOK); 258 hhh->hhh_nhooks--; 259 break; 260 } 261 } 262 HHH_WUNLOCK(hhh); 263 264 return (0); 265 } 266 267 /* 268 * Remove a helper hook function from a helper hook point (including all 269 * virtual instances of the hook point if it is virtualised). 270 */ 271 int 272 hhook_remove_hook_lookup(struct hookinfo *hki) 273 { 274 struct hhook_head *hhh; 275 276 HHHLIST_LOCK(); 277 LIST_FOREACH(hhh, &hhook_head_list, hhh_next) { 278 if (hhh->hhh_type == hki->hook_type && 279 hhh->hhh_id == hki->hook_id) 280 hhook_remove_hook(hhh, hki); 281 } 282 HHHLIST_UNLOCK(); 283 284 return (0); 285 } 286 287 /* 288 * Register a new helper hook point. 289 */ 290 int 291 hhook_head_register(int32_t hhook_type, int32_t hhook_id, struct hhook_head **hhh, 292 uint32_t flags) 293 { 294 struct hhook_head *tmphhh; 295 296 tmphhh = hhook_head_get(hhook_type, hhook_id); 297 298 if (tmphhh != NULL) { 299 /* Hook point previously registered. */ 300 hhook_head_release(tmphhh); 301 return (EEXIST); 302 } 303 304 tmphhh = malloc(sizeof(struct hhook_head), M_HHOOK, 305 M_ZERO | ((flags & HHOOK_WAITOK) ? M_WAITOK : M_NOWAIT)); 306 307 if (tmphhh == NULL) 308 return (ENOMEM); 309 310 tmphhh->hhh_type = hhook_type; 311 tmphhh->hhh_id = hhook_id; 312 tmphhh->hhh_nhooks = 0; 313 STAILQ_INIT(&tmphhh->hhh_hooks); 314 HHH_LOCK_INIT(tmphhh); 315 refcount_init(&tmphhh->hhh_refcount, 1); 316 317 HHHLIST_LOCK(); 318 if (flags & HHOOK_HEADISINVNET) { 319 tmphhh->hhh_flags |= HHH_ISINVNET; 320 #ifdef VIMAGE 321 KASSERT(curvnet != NULL, ("curvnet is NULL")); 322 tmphhh->hhh_vid = (uintptr_t)curvnet; 323 LIST_INSERT_HEAD(&V_hhook_vhead_list, tmphhh, hhh_vnext); 324 #endif 325 } 326 LIST_INSERT_HEAD(&hhook_head_list, tmphhh, hhh_next); 327 n_hhookheads++; 328 HHHLIST_UNLOCK(); 329 330 khelp_new_hhook_registered(tmphhh, flags); 331 332 if (hhh != NULL) 333 *hhh = tmphhh; 334 else 335 refcount_release(&tmphhh->hhh_refcount); 336 337 return (0); 338 } 339 340 static void 341 hhook_head_destroy(struct hhook_head *hhh) 342 { 343 struct hhook *tmp, *tmp2; 344 345 HHHLIST_LOCK_ASSERT(); 346 KASSERT(n_hhookheads > 0, ("n_hhookheads should be > 0")); 347 348 LIST_REMOVE(hhh, hhh_next); 349 #ifdef VIMAGE 350 if (hhook_head_is_virtualised(hhh) == HHOOK_HEADISINVNET) 351 LIST_REMOVE(hhh, hhh_vnext); 352 #endif 353 HHH_WLOCK(hhh); 354 STAILQ_FOREACH_SAFE(tmp, &hhh->hhh_hooks, hhk_next, tmp2) 355 free(tmp, M_HHOOK); 356 HHH_WUNLOCK(hhh); 357 HHH_LOCK_DESTROY(hhh); 358 free(hhh, M_HHOOK); 359 n_hhookheads--; 360 } 361 362 /* 363 * Remove a helper hook point. 364 */ 365 int 366 hhook_head_deregister(struct hhook_head *hhh) 367 { 368 int error; 369 370 error = 0; 371 372 HHHLIST_LOCK(); 373 if (hhh == NULL) 374 error = ENOENT; 375 else if (hhh->hhh_refcount > 1) 376 error = EBUSY; 377 else 378 hhook_head_destroy(hhh); 379 HHHLIST_UNLOCK(); 380 381 return (error); 382 } 383 384 /* 385 * Remove a helper hook point via a hhook_head lookup. 386 */ 387 int 388 hhook_head_deregister_lookup(int32_t hhook_type, int32_t hhook_id) 389 { 390 struct hhook_head *hhh; 391 int error; 392 393 hhh = hhook_head_get(hhook_type, hhook_id); 394 error = hhook_head_deregister(hhh); 395 396 if (error == EBUSY) 397 hhook_head_release(hhh); 398 399 return (error); 400 } 401 402 /* 403 * Lookup and return the hhook_head struct associated with the specified type 404 * and id, or NULL if not found. If found, the hhook_head's refcount is bumped. 405 */ 406 struct hhook_head * 407 hhook_head_get(int32_t hhook_type, int32_t hhook_id) 408 { 409 struct hhook_head *hhh; 410 411 HHHLIST_LOCK(); 412 LIST_FOREACH(hhh, &hhook_head_list, hhh_next) { 413 if (hhh->hhh_type == hhook_type && hhh->hhh_id == hhook_id) { 414 #ifdef VIMAGE 415 if (hhook_head_is_virtualised(hhh) == 416 HHOOK_HEADISINVNET) { 417 KASSERT(curvnet != NULL, ("curvnet is NULL")); 418 if (hhh->hhh_vid != (uintptr_t)curvnet) 419 continue; 420 } 421 #endif 422 refcount_acquire(&hhh->hhh_refcount); 423 break; 424 } 425 } 426 HHHLIST_UNLOCK(); 427 428 return (hhh); 429 } 430 431 void 432 hhook_head_release(struct hhook_head *hhh) 433 { 434 435 refcount_release(&hhh->hhh_refcount); 436 } 437 438 /* 439 * Check the hhook_head private flags and return the appropriate public 440 * representation of the flag to the caller. The function is implemented in a 441 * way that allows us to cope with other subsystems becoming virtualised in the 442 * future. 443 */ 444 uint32_t 445 hhook_head_is_virtualised(struct hhook_head *hhh) 446 { 447 uint32_t ret; 448 449 ret = 0; 450 451 if (hhh != NULL) { 452 if (hhh->hhh_flags & HHH_ISINVNET) 453 ret = HHOOK_HEADISINVNET; 454 } 455 456 return (ret); 457 } 458 459 uint32_t 460 hhook_head_is_virtualised_lookup(int32_t hook_type, int32_t hook_id) 461 { 462 struct hhook_head *hhh; 463 uint32_t ret; 464 465 hhh = hhook_head_get(hook_type, hook_id); 466 467 if (hhh == NULL) 468 return (0); 469 470 ret = hhook_head_is_virtualised(hhh); 471 hhook_head_release(hhh); 472 473 return (ret); 474 } 475 476 /* 477 * Vnet created and being initialised. 478 */ 479 static void 480 hhook_vnet_init(const void *unused __unused) 481 { 482 483 LIST_INIT(&V_hhook_vhead_list); 484 } 485 486 /* 487 * Vnet being torn down and destroyed. 488 */ 489 static void 490 hhook_vnet_uninit(const void *unused __unused) 491 { 492 struct hhook_head *hhh, *tmphhh; 493 494 /* 495 * If subsystems which export helper hook points use the hhook KPI 496 * correctly, the loop below should have no work to do because the 497 * subsystem should have already called hhook_head_deregister(). 498 */ 499 HHHLIST_LOCK(); 500 LIST_FOREACH_SAFE(hhh, &V_hhook_vhead_list, hhh_vnext, tmphhh) { 501 printf("%s: hhook_head type=%d, id=%d cleanup required\n", 502 __func__, hhh->hhh_type, hhh->hhh_id); 503 hhook_head_destroy(hhh); 504 } 505 HHHLIST_UNLOCK(); 506 } 507 508 /* 509 * When a vnet is created and being initialised, init the V_hhook_vhead_list. 510 */ 511 VNET_SYSINIT(hhook_vnet_init, SI_SUB_INIT_IF, SI_ORDER_FIRST, 512 hhook_vnet_init, NULL); 513 514 /* 515 * The hhook KPI provides a mechanism for subsystems which export helper hook 516 * points to clean up on vnet tear down, but in case the KPI is misused, 517 * provide a function to clean up and free memory for a vnet being destroyed. 518 */ 519 VNET_SYSUNINIT(hhook_vnet_uninit, SI_SUB_INIT_IF, SI_ORDER_FIRST, 520 hhook_vnet_uninit, NULL); 521