1b6e66be2SVincenzo Maffione /* 2b6e66be2SVincenzo Maffione * Copyright (C) 2016-2018 Vincenzo Maffione 3b6e66be2SVincenzo Maffione * Copyright (C) 2015 Stefano Garzarella 4b6e66be2SVincenzo Maffione * All rights reserved. 5b6e66be2SVincenzo Maffione * 6b6e66be2SVincenzo Maffione * Redistribution and use in source and binary forms, with or without 7b6e66be2SVincenzo Maffione * modification, are permitted provided that the following conditions 8b6e66be2SVincenzo Maffione * are met: 9b6e66be2SVincenzo Maffione * 1. Redistributions of source code must retain the above copyright 10b6e66be2SVincenzo Maffione * notice, this list of conditions and the following disclaimer. 11b6e66be2SVincenzo Maffione * 2. Redistributions in binary form must reproduce the above copyright 12b6e66be2SVincenzo Maffione * notice, this list of conditions and the following disclaimer in the 13b6e66be2SVincenzo Maffione * documentation and/or other materials provided with the distribution. 14b6e66be2SVincenzo Maffione * 15b6e66be2SVincenzo Maffione * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 16b6e66be2SVincenzo Maffione * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17b6e66be2SVincenzo Maffione * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18b6e66be2SVincenzo Maffione * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 19b6e66be2SVincenzo Maffione * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 20b6e66be2SVincenzo Maffione * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 21b6e66be2SVincenzo Maffione * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 22b6e66be2SVincenzo Maffione * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 23b6e66be2SVincenzo Maffione * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 24b6e66be2SVincenzo Maffione * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 25b6e66be2SVincenzo Maffione * SUCH DAMAGE. 26b6e66be2SVincenzo Maffione * 27b6e66be2SVincenzo Maffione * $FreeBSD$ 28b6e66be2SVincenzo Maffione */ 29b6e66be2SVincenzo Maffione 30b6e66be2SVincenzo Maffione /* 31b6e66be2SVincenzo Maffione * common headers 32b6e66be2SVincenzo Maffione */ 33b6e66be2SVincenzo Maffione #if defined(__FreeBSD__) 34b6e66be2SVincenzo Maffione #include <sys/cdefs.h> 35b6e66be2SVincenzo Maffione #include <sys/param.h> 36b6e66be2SVincenzo Maffione #include <sys/kernel.h> 37b6e66be2SVincenzo Maffione #include <sys/types.h> 38b6e66be2SVincenzo Maffione #include <sys/selinfo.h> 39b6e66be2SVincenzo Maffione #include <sys/socket.h> 40b6e66be2SVincenzo Maffione #include <net/if.h> 41b6e66be2SVincenzo Maffione #include <net/if_var.h> 42b6e66be2SVincenzo Maffione #include <machine/bus.h> 43b6e66be2SVincenzo Maffione 44b6e66be2SVincenzo Maffione #define usleep_range(_1, _2) \ 45b6e66be2SVincenzo Maffione pause_sbt("sync-kloop-sleep", SBT_1US * _1, SBT_1US * 1, C_ABSOLUTE) 46b6e66be2SVincenzo Maffione 47b6e66be2SVincenzo Maffione #elif defined(linux) 48b6e66be2SVincenzo Maffione #include <bsd_glue.h> 49b6e66be2SVincenzo Maffione #include <linux/file.h> 50b6e66be2SVincenzo Maffione #include <linux/eventfd.h> 51b6e66be2SVincenzo Maffione #endif 52b6e66be2SVincenzo Maffione 53b6e66be2SVincenzo Maffione #include <net/netmap.h> 54b6e66be2SVincenzo Maffione #include <dev/netmap/netmap_kern.h> 55b6e66be2SVincenzo Maffione #include <net/netmap_virt.h> 56b6e66be2SVincenzo Maffione #include <dev/netmap/netmap_mem2.h> 57b6e66be2SVincenzo Maffione 58b6e66be2SVincenzo Maffione /* Support for eventfd-based notifications. */ 59b6e66be2SVincenzo Maffione #if defined(linux) 60b6e66be2SVincenzo Maffione #define SYNC_KLOOP_POLL 61b6e66be2SVincenzo Maffione #endif 62b6e66be2SVincenzo Maffione 63b6e66be2SVincenzo Maffione /* Write kring pointers (hwcur, hwtail) to the CSB. 64b6e66be2SVincenzo Maffione * This routine is coupled with ptnetmap_guest_read_kring_csb(). */ 65b6e66be2SVincenzo Maffione static inline void 66b6e66be2SVincenzo Maffione sync_kloop_kernel_write(struct nm_csb_ktoa __user *ptr, uint32_t hwcur, 67b6e66be2SVincenzo Maffione uint32_t hwtail) 68b6e66be2SVincenzo Maffione { 69f79ba6d7SVincenzo Maffione /* Issue a first store-store barrier to make sure writes to the 70f79ba6d7SVincenzo Maffione * netmap ring do not overcome updates on ktoa->hwcur and ktoa->hwtail. */ 71f79ba6d7SVincenzo Maffione nm_stst_barrier(); 72f79ba6d7SVincenzo Maffione 73b6e66be2SVincenzo Maffione /* 74f79ba6d7SVincenzo Maffione * The same scheme used in nm_sync_kloop_appl_write() applies here. 75b6e66be2SVincenzo Maffione * We allow the application to read a value of hwcur more recent than the value 76b6e66be2SVincenzo Maffione * of hwtail, since this would anyway result in a consistent view of the 77b6e66be2SVincenzo Maffione * ring state (and hwcur can never wraparound hwtail, since hwcur must be 78b6e66be2SVincenzo Maffione * behind head). 79b6e66be2SVincenzo Maffione * 80b6e66be2SVincenzo Maffione * The following memory barrier scheme is used to make this happen: 81b6e66be2SVincenzo Maffione * 82b6e66be2SVincenzo Maffione * Application Kernel 83b6e66be2SVincenzo Maffione * 84b6e66be2SVincenzo Maffione * STORE(hwcur) LOAD(hwtail) 85f79ba6d7SVincenzo Maffione * wmb() <-------------> rmb() 86b6e66be2SVincenzo Maffione * STORE(hwtail) LOAD(hwcur) 87b6e66be2SVincenzo Maffione */ 88b6e66be2SVincenzo Maffione CSB_WRITE(ptr, hwcur, hwcur); 89b6e66be2SVincenzo Maffione nm_stst_barrier(); 90b6e66be2SVincenzo Maffione CSB_WRITE(ptr, hwtail, hwtail); 91b6e66be2SVincenzo Maffione } 92b6e66be2SVincenzo Maffione 93b6e66be2SVincenzo Maffione /* Read kring pointers (head, cur, sync_flags) from the CSB. 94b6e66be2SVincenzo Maffione * This routine is coupled with ptnetmap_guest_write_kring_csb(). */ 95b6e66be2SVincenzo Maffione static inline void 96b6e66be2SVincenzo Maffione sync_kloop_kernel_read(struct nm_csb_atok __user *ptr, 97b6e66be2SVincenzo Maffione struct netmap_ring *shadow_ring, 98b6e66be2SVincenzo Maffione uint32_t num_slots) 99b6e66be2SVincenzo Maffione { 100b6e66be2SVincenzo Maffione /* 101b6e66be2SVincenzo Maffione * We place a memory barrier to make sure that the update of head never 102b6e66be2SVincenzo Maffione * overtakes the update of cur. 103f79ba6d7SVincenzo Maffione * (see explanation in sync_kloop_kernel_write). 104b6e66be2SVincenzo Maffione */ 105b6e66be2SVincenzo Maffione CSB_READ(ptr, head, shadow_ring->head); 106f79ba6d7SVincenzo Maffione nm_ldld_barrier(); 107b6e66be2SVincenzo Maffione CSB_READ(ptr, cur, shadow_ring->cur); 108b6e66be2SVincenzo Maffione CSB_READ(ptr, sync_flags, shadow_ring->flags); 109f79ba6d7SVincenzo Maffione 110f79ba6d7SVincenzo Maffione /* Make sure that loads from atok->head and atok->cur are not delayed 111f79ba6d7SVincenzo Maffione * after the loads from the netmap ring. */ 112f79ba6d7SVincenzo Maffione nm_ldld_barrier(); 113b6e66be2SVincenzo Maffione } 114b6e66be2SVincenzo Maffione 115b6e66be2SVincenzo Maffione /* Enable or disable application --> kernel kicks. */ 116b6e66be2SVincenzo Maffione static inline void 117b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(struct nm_csb_ktoa __user *csb_ktoa, uint32_t val) 118b6e66be2SVincenzo Maffione { 119b6e66be2SVincenzo Maffione CSB_WRITE(csb_ktoa, kern_need_kick, val); 120b6e66be2SVincenzo Maffione } 121b6e66be2SVincenzo Maffione 122dde885deSVincenzo Maffione #ifdef SYNC_KLOOP_POLL 123b6e66be2SVincenzo Maffione /* Are application interrupt enabled or disabled? */ 124b6e66be2SVincenzo Maffione static inline uint32_t 125b6e66be2SVincenzo Maffione csb_atok_intr_enabled(struct nm_csb_atok __user *csb_atok) 126b6e66be2SVincenzo Maffione { 127b6e66be2SVincenzo Maffione uint32_t v; 128b6e66be2SVincenzo Maffione 129b6e66be2SVincenzo Maffione CSB_READ(csb_atok, appl_need_kick, v); 130b6e66be2SVincenzo Maffione 131b6e66be2SVincenzo Maffione return v; 132b6e66be2SVincenzo Maffione } 133dde885deSVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 134b6e66be2SVincenzo Maffione 135b6e66be2SVincenzo Maffione static inline void 136b6e66be2SVincenzo Maffione sync_kloop_kring_dump(const char *title, const struct netmap_kring *kring) 137b6e66be2SVincenzo Maffione { 138f79ba6d7SVincenzo Maffione nm_prinf("%s, kring %s, hwcur %d, rhead %d, " 139f79ba6d7SVincenzo Maffione "rcur %d, rtail %d, hwtail %d", 140f79ba6d7SVincenzo Maffione title, kring->name, kring->nr_hwcur, kring->rhead, 141f79ba6d7SVincenzo Maffione kring->rcur, kring->rtail, kring->nr_hwtail); 142b6e66be2SVincenzo Maffione } 143b6e66be2SVincenzo Maffione 144*5faab778SVincenzo Maffione /* Arguments for netmap_sync_kloop_tx_ring() and 145*5faab778SVincenzo Maffione * netmap_sync_kloop_rx_ring(). 146*5faab778SVincenzo Maffione */ 147b6e66be2SVincenzo Maffione struct sync_kloop_ring_args { 148b6e66be2SVincenzo Maffione struct netmap_kring *kring; 149b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok; 150b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa; 151b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 152b6e66be2SVincenzo Maffione struct eventfd_ctx *irq_ctx; 153b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 154*5faab778SVincenzo Maffione /* Are we busy waiting rather than using a schedule() loop ? */ 155*5faab778SVincenzo Maffione bool busy_wait; 156*5faab778SVincenzo Maffione /* Are we processing in the context of VM exit ? */ 157*5faab778SVincenzo Maffione bool direct; 158b6e66be2SVincenzo Maffione }; 159b6e66be2SVincenzo Maffione 160b6e66be2SVincenzo Maffione static void 161b6e66be2SVincenzo Maffione netmap_sync_kloop_tx_ring(const struct sync_kloop_ring_args *a) 162b6e66be2SVincenzo Maffione { 163b6e66be2SVincenzo Maffione struct netmap_kring *kring = a->kring; 164b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok = a->csb_atok; 165b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa = a->csb_ktoa; 166b6e66be2SVincenzo Maffione struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */ 167b6e66be2SVincenzo Maffione bool more_txspace = false; 168b6e66be2SVincenzo Maffione uint32_t num_slots; 169b6e66be2SVincenzo Maffione int batch; 170b6e66be2SVincenzo Maffione 171*5faab778SVincenzo Maffione if (unlikely(nm_kr_tryget(kring, 1, NULL))) { 172*5faab778SVincenzo Maffione return; 173*5faab778SVincenzo Maffione } 174*5faab778SVincenzo Maffione 175b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 176b6e66be2SVincenzo Maffione 177b6e66be2SVincenzo Maffione /* Disable application --> kernel notifications. */ 178*5faab778SVincenzo Maffione if (!a->direct) { 179b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 180*5faab778SVincenzo Maffione } 181b6e66be2SVincenzo Maffione /* Copy the application kring pointers from the CSB */ 182b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 183b6e66be2SVincenzo Maffione 184b6e66be2SVincenzo Maffione for (;;) { 185b6e66be2SVincenzo Maffione batch = shadow_ring.head - kring->nr_hwcur; 186b6e66be2SVincenzo Maffione if (batch < 0) 187b6e66be2SVincenzo Maffione batch += num_slots; 188b6e66be2SVincenzo Maffione 189b6e66be2SVincenzo Maffione #ifdef PTN_TX_BATCH_LIM 190b6e66be2SVincenzo Maffione if (batch > PTN_TX_BATCH_LIM(num_slots)) { 191b6e66be2SVincenzo Maffione /* If application moves ahead too fast, let's cut the move so 192b6e66be2SVincenzo Maffione * that we don't exceed our batch limit. */ 193b6e66be2SVincenzo Maffione uint32_t head_lim = kring->nr_hwcur + PTN_TX_BATCH_LIM(num_slots); 194b6e66be2SVincenzo Maffione 195b6e66be2SVincenzo Maffione if (head_lim >= num_slots) 196b6e66be2SVincenzo Maffione head_lim -= num_slots; 197b6e66be2SVincenzo Maffione nm_prdis(1, "batch: %d head: %d head_lim: %d", batch, shadow_ring.head, 198b6e66be2SVincenzo Maffione head_lim); 199b6e66be2SVincenzo Maffione shadow_ring.head = head_lim; 200b6e66be2SVincenzo Maffione batch = PTN_TX_BATCH_LIM(num_slots); 201b6e66be2SVincenzo Maffione } 202b6e66be2SVincenzo Maffione #endif /* PTN_TX_BATCH_LIM */ 203b6e66be2SVincenzo Maffione 204b6e66be2SVincenzo Maffione if (nm_kr_txspace(kring) <= (num_slots >> 1)) { 205b6e66be2SVincenzo Maffione shadow_ring.flags |= NAF_FORCE_RECLAIM; 206b6e66be2SVincenzo Maffione } 207b6e66be2SVincenzo Maffione 208b6e66be2SVincenzo Maffione /* Netmap prologue */ 209b6e66be2SVincenzo Maffione shadow_ring.tail = kring->rtail; 210b6e66be2SVincenzo Maffione if (unlikely(nm_txsync_prologue(kring, &shadow_ring) >= num_slots)) { 211b6e66be2SVincenzo Maffione /* Reinit ring and enable notifications. */ 212b6e66be2SVincenzo Maffione netmap_ring_reinit(kring); 213*5faab778SVincenzo Maffione if (!a->busy_wait) { 214b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 215*5faab778SVincenzo Maffione } 216b6e66be2SVincenzo Maffione break; 217b6e66be2SVincenzo Maffione } 218b6e66be2SVincenzo Maffione 219b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_TXSYNC)) { 220b6e66be2SVincenzo Maffione sync_kloop_kring_dump("pre txsync", kring); 221b6e66be2SVincenzo Maffione } 222b6e66be2SVincenzo Maffione 223b6e66be2SVincenzo Maffione if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) { 224*5faab778SVincenzo Maffione if (!a->busy_wait) { 225b6e66be2SVincenzo Maffione /* Reenable notifications. */ 226b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 227*5faab778SVincenzo Maffione } 228b6e66be2SVincenzo Maffione nm_prerr("txsync() failed"); 229b6e66be2SVincenzo Maffione break; 230b6e66be2SVincenzo Maffione } 231b6e66be2SVincenzo Maffione 232b6e66be2SVincenzo Maffione /* 233b6e66be2SVincenzo Maffione * Finalize 234b6e66be2SVincenzo Maffione * Copy kernel hwcur and hwtail into the CSB for the application sync(), and 235b6e66be2SVincenzo Maffione * do the nm_sync_finalize. 236b6e66be2SVincenzo Maffione */ 237b6e66be2SVincenzo Maffione sync_kloop_kernel_write(csb_ktoa, kring->nr_hwcur, 238b6e66be2SVincenzo Maffione kring->nr_hwtail); 239b6e66be2SVincenzo Maffione if (kring->rtail != kring->nr_hwtail) { 240b6e66be2SVincenzo Maffione /* Some more room available in the parent adapter. */ 241b6e66be2SVincenzo Maffione kring->rtail = kring->nr_hwtail; 242b6e66be2SVincenzo Maffione more_txspace = true; 243b6e66be2SVincenzo Maffione } 244b6e66be2SVincenzo Maffione 245b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_TXSYNC)) { 246b6e66be2SVincenzo Maffione sync_kloop_kring_dump("post txsync", kring); 247b6e66be2SVincenzo Maffione } 248b6e66be2SVincenzo Maffione 249b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 250b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 251b6e66be2SVincenzo Maffione if (a->irq_ctx && more_txspace && csb_atok_intr_enabled(csb_atok)) { 252*5faab778SVincenzo Maffione /* We could disable kernel --> application kicks here, 253*5faab778SVincenzo Maffione * to avoid spurious interrupts. */ 254b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 255b6e66be2SVincenzo Maffione more_txspace = false; 256b6e66be2SVincenzo Maffione } 257b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 258b6e66be2SVincenzo Maffione 259b6e66be2SVincenzo Maffione /* Read CSB to see if there is more work to do. */ 260b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 261b6e66be2SVincenzo Maffione if (shadow_ring.head == kring->rhead) { 262*5faab778SVincenzo Maffione if (a->busy_wait) { 263*5faab778SVincenzo Maffione break; 264*5faab778SVincenzo Maffione } 265b6e66be2SVincenzo Maffione /* 266b6e66be2SVincenzo Maffione * No more packets to transmit. We enable notifications and 267b6e66be2SVincenzo Maffione * go to sleep, waiting for a kick from the application when new 268b6e66be2SVincenzo Maffione * new slots are ready for transmission. 269b6e66be2SVincenzo Maffione */ 270b6e66be2SVincenzo Maffione /* Reenable notifications. */ 271b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 272f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 273f79ba6d7SVincenzo Maffione nm_stld_barrier(); 274b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 275b6e66be2SVincenzo Maffione if (shadow_ring.head != kring->rhead) { 276b6e66be2SVincenzo Maffione /* We won the race condition, there are more packets to 277b6e66be2SVincenzo Maffione * transmit. Disable notifications and do another cycle */ 278b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 279b6e66be2SVincenzo Maffione continue; 280b6e66be2SVincenzo Maffione } 281b6e66be2SVincenzo Maffione break; 282b6e66be2SVincenzo Maffione } 283b6e66be2SVincenzo Maffione 284b6e66be2SVincenzo Maffione if (nm_kr_txempty(kring)) { 285b6e66be2SVincenzo Maffione /* No more available TX slots. We stop waiting for a notification 286b6e66be2SVincenzo Maffione * from the backend (netmap_tx_irq). */ 287b6e66be2SVincenzo Maffione nm_prdis(1, "TX ring"); 288b6e66be2SVincenzo Maffione break; 289b6e66be2SVincenzo Maffione } 290b6e66be2SVincenzo Maffione } 291b6e66be2SVincenzo Maffione 292*5faab778SVincenzo Maffione nm_kr_put(kring); 293*5faab778SVincenzo Maffione 294b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 295b6e66be2SVincenzo Maffione if (a->irq_ctx && more_txspace && csb_atok_intr_enabled(csb_atok)) { 296b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 297b6e66be2SVincenzo Maffione } 298b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 299b6e66be2SVincenzo Maffione } 300b6e66be2SVincenzo Maffione 301b6e66be2SVincenzo Maffione /* RX cycle without receive any packets */ 302b6e66be2SVincenzo Maffione #define SYNC_LOOP_RX_DRY_CYCLES_MAX 2 303b6e66be2SVincenzo Maffione 304b6e66be2SVincenzo Maffione static inline int 305b6e66be2SVincenzo Maffione sync_kloop_norxslots(struct netmap_kring *kring, uint32_t g_head) 306b6e66be2SVincenzo Maffione { 307b6e66be2SVincenzo Maffione return (NM_ACCESS_ONCE(kring->nr_hwtail) == nm_prev(g_head, 308b6e66be2SVincenzo Maffione kring->nkr_num_slots - 1)); 309b6e66be2SVincenzo Maffione } 310b6e66be2SVincenzo Maffione 311b6e66be2SVincenzo Maffione static void 312b6e66be2SVincenzo Maffione netmap_sync_kloop_rx_ring(const struct sync_kloop_ring_args *a) 313b6e66be2SVincenzo Maffione { 314b6e66be2SVincenzo Maffione 315b6e66be2SVincenzo Maffione struct netmap_kring *kring = a->kring; 316b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok = a->csb_atok; 317b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa = a->csb_ktoa; 318b6e66be2SVincenzo Maffione struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */ 319b6e66be2SVincenzo Maffione int dry_cycles = 0; 320b6e66be2SVincenzo Maffione bool some_recvd = false; 321b6e66be2SVincenzo Maffione uint32_t num_slots; 322b6e66be2SVincenzo Maffione 323*5faab778SVincenzo Maffione if (unlikely(nm_kr_tryget(kring, 1, NULL))) { 324*5faab778SVincenzo Maffione return; 325*5faab778SVincenzo Maffione } 326*5faab778SVincenzo Maffione 327b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 328b6e66be2SVincenzo Maffione 329b6e66be2SVincenzo Maffione /* Get RX csb_atok and csb_ktoa pointers from the CSB. */ 330b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 331b6e66be2SVincenzo Maffione 332b6e66be2SVincenzo Maffione /* Disable notifications. */ 333*5faab778SVincenzo Maffione if (!a->direct) { 334b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 335*5faab778SVincenzo Maffione } 336b6e66be2SVincenzo Maffione /* Copy the application kring pointers from the CSB */ 337b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 338b6e66be2SVincenzo Maffione 339b6e66be2SVincenzo Maffione for (;;) { 340b6e66be2SVincenzo Maffione uint32_t hwtail; 341b6e66be2SVincenzo Maffione 342b6e66be2SVincenzo Maffione /* Netmap prologue */ 343b6e66be2SVincenzo Maffione shadow_ring.tail = kring->rtail; 344b6e66be2SVincenzo Maffione if (unlikely(nm_rxsync_prologue(kring, &shadow_ring) >= num_slots)) { 345b6e66be2SVincenzo Maffione /* Reinit ring and enable notifications. */ 346b6e66be2SVincenzo Maffione netmap_ring_reinit(kring); 347*5faab778SVincenzo Maffione if (!a->busy_wait) { 348b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 349*5faab778SVincenzo Maffione } 350b6e66be2SVincenzo Maffione break; 351b6e66be2SVincenzo Maffione } 352b6e66be2SVincenzo Maffione 353b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_RXSYNC)) { 354b6e66be2SVincenzo Maffione sync_kloop_kring_dump("pre rxsync", kring); 355b6e66be2SVincenzo Maffione } 356b6e66be2SVincenzo Maffione 357b6e66be2SVincenzo Maffione if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) { 358*5faab778SVincenzo Maffione if (!a->busy_wait) { 359b6e66be2SVincenzo Maffione /* Reenable notifications. */ 360b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 361*5faab778SVincenzo Maffione } 362b6e66be2SVincenzo Maffione nm_prerr("rxsync() failed"); 363b6e66be2SVincenzo Maffione break; 364b6e66be2SVincenzo Maffione } 365b6e66be2SVincenzo Maffione 366b6e66be2SVincenzo Maffione /* 367b6e66be2SVincenzo Maffione * Finalize 368b6e66be2SVincenzo Maffione * Copy kernel hwcur and hwtail into the CSB for the application sync() 369b6e66be2SVincenzo Maffione */ 370b6e66be2SVincenzo Maffione hwtail = NM_ACCESS_ONCE(kring->nr_hwtail); 371b6e66be2SVincenzo Maffione sync_kloop_kernel_write(csb_ktoa, kring->nr_hwcur, hwtail); 372b6e66be2SVincenzo Maffione if (kring->rtail != hwtail) { 373b6e66be2SVincenzo Maffione kring->rtail = hwtail; 374b6e66be2SVincenzo Maffione some_recvd = true; 375b6e66be2SVincenzo Maffione dry_cycles = 0; 376b6e66be2SVincenzo Maffione } else { 377b6e66be2SVincenzo Maffione dry_cycles++; 378b6e66be2SVincenzo Maffione } 379b6e66be2SVincenzo Maffione 380b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_RXSYNC)) { 381b6e66be2SVincenzo Maffione sync_kloop_kring_dump("post rxsync", kring); 382b6e66be2SVincenzo Maffione } 383b6e66be2SVincenzo Maffione 384b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 385b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 386b6e66be2SVincenzo Maffione if (a->irq_ctx && some_recvd && csb_atok_intr_enabled(csb_atok)) { 387*5faab778SVincenzo Maffione /* We could disable kernel --> application kicks here, 388*5faab778SVincenzo Maffione * to avoid spurious interrupts. */ 389b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 390b6e66be2SVincenzo Maffione some_recvd = false; 391b6e66be2SVincenzo Maffione } 392b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 393b6e66be2SVincenzo Maffione 394b6e66be2SVincenzo Maffione /* Read CSB to see if there is more work to do. */ 395b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 396b6e66be2SVincenzo Maffione if (sync_kloop_norxslots(kring, shadow_ring.head)) { 397*5faab778SVincenzo Maffione if (a->busy_wait) { 398*5faab778SVincenzo Maffione break; 399*5faab778SVincenzo Maffione } 400b6e66be2SVincenzo Maffione /* 401b6e66be2SVincenzo Maffione * No more slots available for reception. We enable notification and 402b6e66be2SVincenzo Maffione * go to sleep, waiting for a kick from the application when new receive 403b6e66be2SVincenzo Maffione * slots are available. 404b6e66be2SVincenzo Maffione */ 405b6e66be2SVincenzo Maffione /* Reenable notifications. */ 406b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 407f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 408f79ba6d7SVincenzo Maffione nm_stld_barrier(); 409b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 410b6e66be2SVincenzo Maffione if (!sync_kloop_norxslots(kring, shadow_ring.head)) { 411b6e66be2SVincenzo Maffione /* We won the race condition, more slots are available. Disable 412b6e66be2SVincenzo Maffione * notifications and do another cycle. */ 413b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 414b6e66be2SVincenzo Maffione continue; 415b6e66be2SVincenzo Maffione } 416b6e66be2SVincenzo Maffione break; 417b6e66be2SVincenzo Maffione } 418b6e66be2SVincenzo Maffione 419b6e66be2SVincenzo Maffione hwtail = NM_ACCESS_ONCE(kring->nr_hwtail); 420b6e66be2SVincenzo Maffione if (unlikely(hwtail == kring->rhead || 421b6e66be2SVincenzo Maffione dry_cycles >= SYNC_LOOP_RX_DRY_CYCLES_MAX)) { 422b6e66be2SVincenzo Maffione /* No more packets to be read from the backend. We stop and 423b6e66be2SVincenzo Maffione * wait for a notification from the backend (netmap_rx_irq). */ 424b6e66be2SVincenzo Maffione nm_prdis(1, "nr_hwtail: %d rhead: %d dry_cycles: %d", 425b6e66be2SVincenzo Maffione hwtail, kring->rhead, dry_cycles); 426b6e66be2SVincenzo Maffione break; 427b6e66be2SVincenzo Maffione } 428b6e66be2SVincenzo Maffione } 429b6e66be2SVincenzo Maffione 430b6e66be2SVincenzo Maffione nm_kr_put(kring); 431b6e66be2SVincenzo Maffione 432b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 433b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 434b6e66be2SVincenzo Maffione if (a->irq_ctx && some_recvd && csb_atok_intr_enabled(csb_atok)) { 435b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 436b6e66be2SVincenzo Maffione } 437b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 438b6e66be2SVincenzo Maffione } 439b6e66be2SVincenzo Maffione 440b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 441*5faab778SVincenzo Maffione struct sync_kloop_poll_ctx; 442b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry { 443b6e66be2SVincenzo Maffione /* Support for receiving notifications from 444b6e66be2SVincenzo Maffione * a netmap ring or from the application. */ 445b6e66be2SVincenzo Maffione struct file *filp; 446b6e66be2SVincenzo Maffione wait_queue_t wait; 447b6e66be2SVincenzo Maffione wait_queue_head_t *wqh; 448b6e66be2SVincenzo Maffione 449b6e66be2SVincenzo Maffione /* Support for sending notifications to the application. */ 450b6e66be2SVincenzo Maffione struct eventfd_ctx *irq_ctx; 451b6e66be2SVincenzo Maffione struct file *irq_filp; 452*5faab778SVincenzo Maffione 453*5faab778SVincenzo Maffione /* Arguments for the ring processing function. Useful 454*5faab778SVincenzo Maffione * in case of custom wake-up function. */ 455*5faab778SVincenzo Maffione struct sync_kloop_ring_args *args; 456*5faab778SVincenzo Maffione struct sync_kloop_poll_ctx *parent; 457*5faab778SVincenzo Maffione 458b6e66be2SVincenzo Maffione }; 459b6e66be2SVincenzo Maffione 460b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx { 461b6e66be2SVincenzo Maffione poll_table wait_table; 462b6e66be2SVincenzo Maffione unsigned int next_entry; 463*5faab778SVincenzo Maffione int (*next_wake_fun)(wait_queue_t *, unsigned, int, void *); 464b6e66be2SVincenzo Maffione unsigned int num_entries; 465*5faab778SVincenzo Maffione unsigned int num_tx_rings; 466*5faab778SVincenzo Maffione unsigned int num_rings; 467*5faab778SVincenzo Maffione /* First num_tx_rings entries are for the TX kicks. 468*5faab778SVincenzo Maffione * Then the RX kicks entries follow. The last two 469*5faab778SVincenzo Maffione * entries are for TX irq, and RX irq. */ 470b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry entries[0]; 471b6e66be2SVincenzo Maffione }; 472b6e66be2SVincenzo Maffione 473b6e66be2SVincenzo Maffione static void 474b6e66be2SVincenzo Maffione sync_kloop_poll_table_queue_proc(struct file *file, wait_queue_head_t *wqh, 475b6e66be2SVincenzo Maffione poll_table *pt) 476b6e66be2SVincenzo Maffione { 477b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = 478b6e66be2SVincenzo Maffione container_of(pt, struct sync_kloop_poll_ctx, wait_table); 479b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry *entry = poll_ctx->entries + 480b6e66be2SVincenzo Maffione poll_ctx->next_entry; 481b6e66be2SVincenzo Maffione 482b6e66be2SVincenzo Maffione BUG_ON(poll_ctx->next_entry >= poll_ctx->num_entries); 483b6e66be2SVincenzo Maffione entry->wqh = wqh; 484b6e66be2SVincenzo Maffione entry->filp = file; 485b6e66be2SVincenzo Maffione /* Use the default wake up function. */ 486*5faab778SVincenzo Maffione if (poll_ctx->next_wake_fun == NULL) { 487b6e66be2SVincenzo Maffione init_waitqueue_entry(&entry->wait, current); 488*5faab778SVincenzo Maffione } else { 489*5faab778SVincenzo Maffione init_waitqueue_func_entry(&entry->wait, 490*5faab778SVincenzo Maffione poll_ctx->next_wake_fun); 491*5faab778SVincenzo Maffione } 492b6e66be2SVincenzo Maffione add_wait_queue(wqh, &entry->wait); 493*5faab778SVincenzo Maffione } 494*5faab778SVincenzo Maffione 495*5faab778SVincenzo Maffione static int 496*5faab778SVincenzo Maffione sync_kloop_tx_kick_wake_fun(wait_queue_t *wait, unsigned mode, 497*5faab778SVincenzo Maffione int wake_flags, void *key) 498*5faab778SVincenzo Maffione { 499*5faab778SVincenzo Maffione struct sync_kloop_poll_entry *entry = 500*5faab778SVincenzo Maffione container_of(wait, struct sync_kloop_poll_entry, wait); 501*5faab778SVincenzo Maffione 502*5faab778SVincenzo Maffione netmap_sync_kloop_tx_ring(entry->args); 503*5faab778SVincenzo Maffione 504*5faab778SVincenzo Maffione return 0; 505*5faab778SVincenzo Maffione } 506*5faab778SVincenzo Maffione 507*5faab778SVincenzo Maffione static int 508*5faab778SVincenzo Maffione sync_kloop_tx_irq_wake_fun(wait_queue_t *wait, unsigned mode, 509*5faab778SVincenzo Maffione int wake_flags, void *key) 510*5faab778SVincenzo Maffione { 511*5faab778SVincenzo Maffione struct sync_kloop_poll_entry *entry = 512*5faab778SVincenzo Maffione container_of(wait, struct sync_kloop_poll_entry, wait); 513*5faab778SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = entry->parent; 514*5faab778SVincenzo Maffione int i; 515*5faab778SVincenzo Maffione 516*5faab778SVincenzo Maffione for (i = 0; i < poll_ctx->num_tx_rings; i++) { 517*5faab778SVincenzo Maffione struct eventfd_ctx *irq_ctx = poll_ctx->entries[i].irq_ctx; 518*5faab778SVincenzo Maffione 519*5faab778SVincenzo Maffione if (irq_ctx) { 520*5faab778SVincenzo Maffione eventfd_signal(irq_ctx, 1); 521*5faab778SVincenzo Maffione } 522*5faab778SVincenzo Maffione } 523*5faab778SVincenzo Maffione 524*5faab778SVincenzo Maffione return 0; 525*5faab778SVincenzo Maffione } 526*5faab778SVincenzo Maffione 527*5faab778SVincenzo Maffione static int 528*5faab778SVincenzo Maffione sync_kloop_rx_kick_wake_fun(wait_queue_t *wait, unsigned mode, 529*5faab778SVincenzo Maffione int wake_flags, void *key) 530*5faab778SVincenzo Maffione { 531*5faab778SVincenzo Maffione struct sync_kloop_poll_entry *entry = 532*5faab778SVincenzo Maffione container_of(wait, struct sync_kloop_poll_entry, wait); 533*5faab778SVincenzo Maffione 534*5faab778SVincenzo Maffione netmap_sync_kloop_rx_ring(entry->args); 535*5faab778SVincenzo Maffione 536*5faab778SVincenzo Maffione return 0; 537*5faab778SVincenzo Maffione } 538*5faab778SVincenzo Maffione 539*5faab778SVincenzo Maffione static int 540*5faab778SVincenzo Maffione sync_kloop_rx_irq_wake_fun(wait_queue_t *wait, unsigned mode, 541*5faab778SVincenzo Maffione int wake_flags, void *key) 542*5faab778SVincenzo Maffione { 543*5faab778SVincenzo Maffione struct sync_kloop_poll_entry *entry = 544*5faab778SVincenzo Maffione container_of(wait, struct sync_kloop_poll_entry, wait); 545*5faab778SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = entry->parent; 546*5faab778SVincenzo Maffione int i; 547*5faab778SVincenzo Maffione 548*5faab778SVincenzo Maffione for (i = poll_ctx->num_tx_rings; i < poll_ctx->num_rings; i++) { 549*5faab778SVincenzo Maffione struct eventfd_ctx *irq_ctx = poll_ctx->entries[i].irq_ctx; 550*5faab778SVincenzo Maffione 551*5faab778SVincenzo Maffione if (irq_ctx) { 552*5faab778SVincenzo Maffione eventfd_signal(irq_ctx, 1); 553*5faab778SVincenzo Maffione } 554*5faab778SVincenzo Maffione } 555*5faab778SVincenzo Maffione 556*5faab778SVincenzo Maffione return 0; 557b6e66be2SVincenzo Maffione } 558b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 559b6e66be2SVincenzo Maffione 560b6e66be2SVincenzo Maffione int 561b6e66be2SVincenzo Maffione netmap_sync_kloop(struct netmap_priv_d *priv, struct nmreq_header *hdr) 562b6e66be2SVincenzo Maffione { 563b6e66be2SVincenzo Maffione struct nmreq_sync_kloop_start *req = 564b6e66be2SVincenzo Maffione (struct nmreq_sync_kloop_start *)(uintptr_t)hdr->nr_body; 565b6e66be2SVincenzo Maffione struct nmreq_opt_sync_kloop_eventfds *eventfds_opt = NULL; 566b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 567b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = NULL; 568b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 569b6e66be2SVincenzo Maffione int num_rx_rings, num_tx_rings, num_rings; 570f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *args = NULL; 571b6e66be2SVincenzo Maffione uint32_t sleep_us = req->sleep_us; 572b6e66be2SVincenzo Maffione struct nm_csb_atok* csb_atok_base; 573b6e66be2SVincenzo Maffione struct nm_csb_ktoa* csb_ktoa_base; 574b6e66be2SVincenzo Maffione struct netmap_adapter *na; 575b6e66be2SVincenzo Maffione struct nmreq_option *opt; 576*5faab778SVincenzo Maffione bool na_could_sleep = false; 577*5faab778SVincenzo Maffione bool busy_wait = true; 578*5faab778SVincenzo Maffione bool direct_tx = false; 579*5faab778SVincenzo Maffione bool direct_rx = false; 580b6e66be2SVincenzo Maffione int err = 0; 581b6e66be2SVincenzo Maffione int i; 582b6e66be2SVincenzo Maffione 583b6e66be2SVincenzo Maffione if (sleep_us > 1000000) { 584b6e66be2SVincenzo Maffione /* We do not accept sleeping for more than a second. */ 585b6e66be2SVincenzo Maffione return EINVAL; 586b6e66be2SVincenzo Maffione } 587b6e66be2SVincenzo Maffione 588b6e66be2SVincenzo Maffione if (priv->np_nifp == NULL) { 589b6e66be2SVincenzo Maffione return ENXIO; 590b6e66be2SVincenzo Maffione } 591b6e66be2SVincenzo Maffione mb(); /* make sure following reads are not from cache */ 592b6e66be2SVincenzo Maffione 593b6e66be2SVincenzo Maffione na = priv->np_na; 594b6e66be2SVincenzo Maffione if (!nm_netmap_on(na)) { 595b6e66be2SVincenzo Maffione return ENXIO; 596b6e66be2SVincenzo Maffione } 597b6e66be2SVincenzo Maffione 598b6e66be2SVincenzo Maffione NMG_LOCK(); 599b6e66be2SVincenzo Maffione /* Make sure the application is working in CSB mode. */ 600b6e66be2SVincenzo Maffione if (!priv->np_csb_atok_base || !priv->np_csb_ktoa_base) { 601b6e66be2SVincenzo Maffione NMG_UNLOCK(); 602b6e66be2SVincenzo Maffione nm_prerr("sync-kloop on %s requires " 603b6e66be2SVincenzo Maffione "NETMAP_REQ_OPT_CSB option", na->name); 604b6e66be2SVincenzo Maffione return EINVAL; 605b6e66be2SVincenzo Maffione } 606b6e66be2SVincenzo Maffione 607b6e66be2SVincenzo Maffione csb_atok_base = priv->np_csb_atok_base; 608b6e66be2SVincenzo Maffione csb_ktoa_base = priv->np_csb_ktoa_base; 609b6e66be2SVincenzo Maffione 610b6e66be2SVincenzo Maffione /* Make sure that no kloop is currently running. */ 611b6e66be2SVincenzo Maffione if (priv->np_kloop_state & NM_SYNC_KLOOP_RUNNING) { 612b6e66be2SVincenzo Maffione err = EBUSY; 613b6e66be2SVincenzo Maffione } 614b6e66be2SVincenzo Maffione priv->np_kloop_state |= NM_SYNC_KLOOP_RUNNING; 615b6e66be2SVincenzo Maffione NMG_UNLOCK(); 616b6e66be2SVincenzo Maffione if (err) { 617b6e66be2SVincenzo Maffione return err; 618b6e66be2SVincenzo Maffione } 619b6e66be2SVincenzo Maffione 620b6e66be2SVincenzo Maffione num_rx_rings = priv->np_qlast[NR_RX] - priv->np_qfirst[NR_RX]; 621b6e66be2SVincenzo Maffione num_tx_rings = priv->np_qlast[NR_TX] - priv->np_qfirst[NR_TX]; 622b6e66be2SVincenzo Maffione num_rings = num_tx_rings + num_rx_rings; 623b6e66be2SVincenzo Maffione 624f79ba6d7SVincenzo Maffione args = nm_os_malloc(num_rings * sizeof(args[0])); 625f79ba6d7SVincenzo Maffione if (!args) { 626f79ba6d7SVincenzo Maffione err = ENOMEM; 627f79ba6d7SVincenzo Maffione goto out; 628f79ba6d7SVincenzo Maffione } 629f79ba6d7SVincenzo Maffione 630*5faab778SVincenzo Maffione /* Prepare the arguments for netmap_sync_kloop_tx_ring() 631*5faab778SVincenzo Maffione * and netmap_sync_kloop_rx_ring(). */ 632*5faab778SVincenzo Maffione for (i = 0; i < num_tx_rings; i++) { 633*5faab778SVincenzo Maffione struct sync_kloop_ring_args *a = args + i; 634*5faab778SVincenzo Maffione 635*5faab778SVincenzo Maffione a->kring = NMR(na, NR_TX)[i + priv->np_qfirst[NR_TX]]; 636*5faab778SVincenzo Maffione a->csb_atok = csb_atok_base + i; 637*5faab778SVincenzo Maffione a->csb_ktoa = csb_ktoa_base + i; 638*5faab778SVincenzo Maffione a->busy_wait = busy_wait; 639*5faab778SVincenzo Maffione a->direct = direct_tx; 640*5faab778SVincenzo Maffione } 641*5faab778SVincenzo Maffione for (i = 0; i < num_rx_rings; i++) { 642*5faab778SVincenzo Maffione struct sync_kloop_ring_args *a = args + num_tx_rings + i; 643*5faab778SVincenzo Maffione 644*5faab778SVincenzo Maffione a->kring = NMR(na, NR_RX)[i + priv->np_qfirst[NR_RX]]; 645*5faab778SVincenzo Maffione a->csb_atok = csb_atok_base + num_tx_rings + i; 646*5faab778SVincenzo Maffione a->csb_ktoa = csb_ktoa_base + num_tx_rings + i; 647*5faab778SVincenzo Maffione a->busy_wait = busy_wait; 648*5faab778SVincenzo Maffione a->direct = direct_rx; 649*5faab778SVincenzo Maffione } 650*5faab778SVincenzo Maffione 651b6e66be2SVincenzo Maffione /* Validate notification options. */ 652b6e66be2SVincenzo Maffione opt = nmreq_findoption((struct nmreq_option *)(uintptr_t)hdr->nr_options, 653*5faab778SVincenzo Maffione NETMAP_REQ_OPT_SYNC_KLOOP_MODE); 654*5faab778SVincenzo Maffione if (opt != NULL) { 655*5faab778SVincenzo Maffione struct nmreq_opt_sync_kloop_mode *mode_opt = 656*5faab778SVincenzo Maffione (struct nmreq_opt_sync_kloop_mode *)opt; 657*5faab778SVincenzo Maffione 658*5faab778SVincenzo Maffione direct_tx = !!(mode_opt->mode & NM_OPT_SYNC_KLOOP_DIRECT_TX); 659*5faab778SVincenzo Maffione direct_rx = !!(mode_opt->mode & NM_OPT_SYNC_KLOOP_DIRECT_RX); 660*5faab778SVincenzo Maffione if (mode_opt->mode & ~(NM_OPT_SYNC_KLOOP_DIRECT_TX | 661*5faab778SVincenzo Maffione NM_OPT_SYNC_KLOOP_DIRECT_RX)) { 662*5faab778SVincenzo Maffione opt->nro_status = err = EINVAL; 663*5faab778SVincenzo Maffione goto out; 664*5faab778SVincenzo Maffione } 665*5faab778SVincenzo Maffione opt->nro_status = 0; 666*5faab778SVincenzo Maffione } 667*5faab778SVincenzo Maffione opt = nmreq_findoption((struct nmreq_option *)(uintptr_t)hdr->nr_options, 668b6e66be2SVincenzo Maffione NETMAP_REQ_OPT_SYNC_KLOOP_EVENTFDS); 669b6e66be2SVincenzo Maffione if (opt != NULL) { 670b6e66be2SVincenzo Maffione err = nmreq_checkduplicate(opt); 671b6e66be2SVincenzo Maffione if (err) { 672b6e66be2SVincenzo Maffione opt->nro_status = err; 673b6e66be2SVincenzo Maffione goto out; 674b6e66be2SVincenzo Maffione } 675b6e66be2SVincenzo Maffione if (opt->nro_size != sizeof(*eventfds_opt) + 676b6e66be2SVincenzo Maffione sizeof(eventfds_opt->eventfds[0]) * num_rings) { 677b6e66be2SVincenzo Maffione /* Option size not consistent with the number of 678b6e66be2SVincenzo Maffione * entries. */ 679b6e66be2SVincenzo Maffione opt->nro_status = err = EINVAL; 680b6e66be2SVincenzo Maffione goto out; 681b6e66be2SVincenzo Maffione } 682b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 683b6e66be2SVincenzo Maffione eventfds_opt = (struct nmreq_opt_sync_kloop_eventfds *)opt; 684b6e66be2SVincenzo Maffione opt->nro_status = 0; 685*5faab778SVincenzo Maffione 686*5faab778SVincenzo Maffione /* Check if some ioeventfd entry is not defined, and force sleep 687*5faab778SVincenzo Maffione * synchronization in that case. */ 688*5faab778SVincenzo Maffione busy_wait = false; 689*5faab778SVincenzo Maffione for (i = 0; i < num_rings; i++) { 690*5faab778SVincenzo Maffione if (eventfds_opt->eventfds[i].ioeventfd < 0) { 691*5faab778SVincenzo Maffione busy_wait = true; 692*5faab778SVincenzo Maffione break; 693*5faab778SVincenzo Maffione } 694*5faab778SVincenzo Maffione } 695*5faab778SVincenzo Maffione 696*5faab778SVincenzo Maffione if (busy_wait && (direct_tx || direct_rx)) { 697*5faab778SVincenzo Maffione /* For direct processing we need all the 698*5faab778SVincenzo Maffione * ioeventfds to be valid. */ 699*5faab778SVincenzo Maffione opt->nro_status = err = EINVAL; 700*5faab778SVincenzo Maffione goto out; 701*5faab778SVincenzo Maffione } 702*5faab778SVincenzo Maffione 703b6e66be2SVincenzo Maffione /* We need 2 poll entries for TX and RX notifications coming 704b6e66be2SVincenzo Maffione * from the netmap adapter, plus one entries per ring for the 705b6e66be2SVincenzo Maffione * notifications coming from the application. */ 706b6e66be2SVincenzo Maffione poll_ctx = nm_os_malloc(sizeof(*poll_ctx) + 707*5faab778SVincenzo Maffione (num_rings + 2) * sizeof(poll_ctx->entries[0])); 708b6e66be2SVincenzo Maffione init_poll_funcptr(&poll_ctx->wait_table, 709b6e66be2SVincenzo Maffione sync_kloop_poll_table_queue_proc); 710b6e66be2SVincenzo Maffione poll_ctx->num_entries = 2 + num_rings; 711*5faab778SVincenzo Maffione poll_ctx->num_tx_rings = num_tx_rings; 712*5faab778SVincenzo Maffione poll_ctx->num_rings = num_rings; 713b6e66be2SVincenzo Maffione poll_ctx->next_entry = 0; 714*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = NULL; 715*5faab778SVincenzo Maffione 716*5faab778SVincenzo Maffione if (direct_tx && (na->na_flags & NAF_BDG_MAYSLEEP)) { 717*5faab778SVincenzo Maffione /* In direct mode, VALE txsync is called from 718*5faab778SVincenzo Maffione * wake-up context, where it is not possible 719*5faab778SVincenzo Maffione * to sleep. 720*5faab778SVincenzo Maffione */ 721*5faab778SVincenzo Maffione na->na_flags &= ~NAF_BDG_MAYSLEEP; 722*5faab778SVincenzo Maffione na_could_sleep = true; 723*5faab778SVincenzo Maffione } 724*5faab778SVincenzo Maffione 725*5faab778SVincenzo Maffione for (i = 0; i < num_rings + 2; i++) { 726*5faab778SVincenzo Maffione poll_ctx->entries[i].args = args + i; 727*5faab778SVincenzo Maffione poll_ctx->entries[i].parent = poll_ctx; 728*5faab778SVincenzo Maffione } 729*5faab778SVincenzo Maffione 730b6e66be2SVincenzo Maffione /* Poll for notifications coming from the applications through 731b6e66be2SVincenzo Maffione * eventfds. */ 732*5faab778SVincenzo Maffione for (i = 0; i < num_rings; i++, poll_ctx->next_entry++) { 733*5faab778SVincenzo Maffione struct eventfd_ctx *irq = NULL; 734*5faab778SVincenzo Maffione struct file *filp = NULL; 735b6e66be2SVincenzo Maffione unsigned long mask; 736*5faab778SVincenzo Maffione bool tx_ring = (i < num_tx_rings); 737b6e66be2SVincenzo Maffione 738*5faab778SVincenzo Maffione if (eventfds_opt->eventfds[i].irqfd >= 0) { 739*5faab778SVincenzo Maffione filp = eventfd_fget( 740*5faab778SVincenzo Maffione eventfds_opt->eventfds[i].irqfd); 741b6e66be2SVincenzo Maffione if (IS_ERR(filp)) { 742b6e66be2SVincenzo Maffione err = PTR_ERR(filp); 743b6e66be2SVincenzo Maffione goto out; 744b6e66be2SVincenzo Maffione } 745b6e66be2SVincenzo Maffione irq = eventfd_ctx_fileget(filp); 746b6e66be2SVincenzo Maffione if (IS_ERR(irq)) { 747b6e66be2SVincenzo Maffione err = PTR_ERR(irq); 748b6e66be2SVincenzo Maffione goto out; 749b6e66be2SVincenzo Maffione } 750b6e66be2SVincenzo Maffione } 751*5faab778SVincenzo Maffione poll_ctx->entries[i].irq_filp = filp; 752*5faab778SVincenzo Maffione poll_ctx->entries[i].irq_ctx = irq; 753*5faab778SVincenzo Maffione poll_ctx->entries[i].args->busy_wait = busy_wait; 754*5faab778SVincenzo Maffione /* Don't let netmap_sync_kloop_*x_ring() use 755*5faab778SVincenzo Maffione * IRQs in direct mode. */ 756*5faab778SVincenzo Maffione poll_ctx->entries[i].args->irq_ctx = 757*5faab778SVincenzo Maffione ((tx_ring && direct_tx) || 758*5faab778SVincenzo Maffione (!tx_ring && direct_rx)) ? NULL : 759*5faab778SVincenzo Maffione poll_ctx->entries[i].irq_ctx; 760*5faab778SVincenzo Maffione poll_ctx->entries[i].args->direct = 761*5faab778SVincenzo Maffione (tx_ring ? direct_tx : direct_rx); 762*5faab778SVincenzo Maffione 763*5faab778SVincenzo Maffione if (!busy_wait) { 764*5faab778SVincenzo Maffione filp = eventfd_fget( 765*5faab778SVincenzo Maffione eventfds_opt->eventfds[i].ioeventfd); 766*5faab778SVincenzo Maffione if (IS_ERR(filp)) { 767*5faab778SVincenzo Maffione err = PTR_ERR(filp); 768*5faab778SVincenzo Maffione goto out; 769*5faab778SVincenzo Maffione } 770*5faab778SVincenzo Maffione if (tx_ring && direct_tx) { 771*5faab778SVincenzo Maffione /* Override the wake up function 772*5faab778SVincenzo Maffione * so that it can directly call 773*5faab778SVincenzo Maffione * netmap_sync_kloop_tx_ring(). 774*5faab778SVincenzo Maffione */ 775*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = 776*5faab778SVincenzo Maffione sync_kloop_tx_kick_wake_fun; 777*5faab778SVincenzo Maffione } else if (!tx_ring && direct_rx) { 778*5faab778SVincenzo Maffione /* Same for direct RX. */ 779*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = 780*5faab778SVincenzo Maffione sync_kloop_rx_kick_wake_fun; 781*5faab778SVincenzo Maffione } else { 782*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = NULL; 783*5faab778SVincenzo Maffione } 784*5faab778SVincenzo Maffione mask = filp->f_op->poll(filp, 785*5faab778SVincenzo Maffione &poll_ctx->wait_table); 786*5faab778SVincenzo Maffione if (mask & POLLERR) { 787*5faab778SVincenzo Maffione err = EINVAL; 788*5faab778SVincenzo Maffione goto out; 789*5faab778SVincenzo Maffione } 790*5faab778SVincenzo Maffione } 791*5faab778SVincenzo Maffione } 792*5faab778SVincenzo Maffione 793b6e66be2SVincenzo Maffione /* Poll for notifications coming from the netmap rings bound to 794b6e66be2SVincenzo Maffione * this file descriptor. */ 795*5faab778SVincenzo Maffione if (!busy_wait) { 796b6e66be2SVincenzo Maffione NMG_LOCK(); 797*5faab778SVincenzo Maffione /* In direct mode, override the wake up function so 798*5faab778SVincenzo Maffione * that it can forward the netmap_tx_irq() to the 799*5faab778SVincenzo Maffione * guest. */ 800*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = direct_tx ? 801*5faab778SVincenzo Maffione sync_kloop_tx_irq_wake_fun : NULL; 802a56136a1SVincenzo Maffione poll_wait(priv->np_filp, priv->np_si[NR_TX], 803a56136a1SVincenzo Maffione &poll_ctx->wait_table); 804*5faab778SVincenzo Maffione poll_ctx->next_entry++; 805*5faab778SVincenzo Maffione 806*5faab778SVincenzo Maffione poll_ctx->next_wake_fun = direct_rx ? 807*5faab778SVincenzo Maffione sync_kloop_rx_irq_wake_fun : NULL; 808a56136a1SVincenzo Maffione poll_wait(priv->np_filp, priv->np_si[NR_RX], 809a56136a1SVincenzo Maffione &poll_ctx->wait_table); 810*5faab778SVincenzo Maffione poll_ctx->next_entry++; 811b6e66be2SVincenzo Maffione NMG_UNLOCK(); 812b6e66be2SVincenzo Maffione } 813b6e66be2SVincenzo Maffione #else /* SYNC_KLOOP_POLL */ 814b6e66be2SVincenzo Maffione opt->nro_status = EOPNOTSUPP; 815b6e66be2SVincenzo Maffione goto out; 816b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 817b6e66be2SVincenzo Maffione } 818b6e66be2SVincenzo Maffione 819*5faab778SVincenzo Maffione nm_prinf("kloop busy_wait %u, direct_tx %u, direct_rx %u, " 820*5faab778SVincenzo Maffione "na_could_sleep %u", busy_wait, direct_tx, direct_rx, 821*5faab778SVincenzo Maffione na_could_sleep); 822f79ba6d7SVincenzo Maffione 823b6e66be2SVincenzo Maffione /* Main loop. */ 824b6e66be2SVincenzo Maffione for (;;) { 825b6e66be2SVincenzo Maffione if (unlikely(NM_ACCESS_ONCE(priv->np_kloop_state) & NM_SYNC_KLOOP_STOPPING)) { 826b6e66be2SVincenzo Maffione break; 827b6e66be2SVincenzo Maffione } 828b6e66be2SVincenzo Maffione 829b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 830*5faab778SVincenzo Maffione if (!busy_wait) { 831f79ba6d7SVincenzo Maffione /* It is important to set the task state as 832f79ba6d7SVincenzo Maffione * interruptible before processing any TX/RX ring, 833f79ba6d7SVincenzo Maffione * so that if a notification on ring Y comes after 834f79ba6d7SVincenzo Maffione * we have processed ring Y, but before we call 835f79ba6d7SVincenzo Maffione * schedule(), we don't miss it. This is true because 836f79ba6d7SVincenzo Maffione * the wake up function will change the the task state, 837f79ba6d7SVincenzo Maffione * and therefore the schedule_timeout() call below 838f79ba6d7SVincenzo Maffione * will observe the change). 839f79ba6d7SVincenzo Maffione */ 840f79ba6d7SVincenzo Maffione set_current_state(TASK_INTERRUPTIBLE); 841f79ba6d7SVincenzo Maffione } 842b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 843b6e66be2SVincenzo Maffione 844b6e66be2SVincenzo Maffione /* Process all the TX rings bound to this file descriptor. */ 845*5faab778SVincenzo Maffione for (i = 0; !direct_tx && i < num_tx_rings; i++) { 846f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + i; 847f79ba6d7SVincenzo Maffione netmap_sync_kloop_tx_ring(a); 848b6e66be2SVincenzo Maffione } 849b6e66be2SVincenzo Maffione 850b6e66be2SVincenzo Maffione /* Process all the RX rings bound to this file descriptor. */ 851*5faab778SVincenzo Maffione for (i = 0; !direct_rx && i < num_rx_rings; i++) { 852f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + num_tx_rings + i; 853f79ba6d7SVincenzo Maffione netmap_sync_kloop_rx_ring(a); 854b6e66be2SVincenzo Maffione } 855b6e66be2SVincenzo Maffione 856*5faab778SVincenzo Maffione if (busy_wait) { 857b6e66be2SVincenzo Maffione /* Default synchronization method: sleep for a while. */ 858b6e66be2SVincenzo Maffione usleep_range(sleep_us, sleep_us); 859b6e66be2SVincenzo Maffione } 860*5faab778SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 861*5faab778SVincenzo Maffione else { 862*5faab778SVincenzo Maffione /* Yield to the scheduler waiting for a notification 863*5faab778SVincenzo Maffione * to come either from netmap or the application. */ 864*5faab778SVincenzo Maffione schedule_timeout(msecs_to_jiffies(3000)); 865*5faab778SVincenzo Maffione } 866*5faab778SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 867b6e66be2SVincenzo Maffione } 868b6e66be2SVincenzo Maffione out: 869b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 870b6e66be2SVincenzo Maffione if (poll_ctx) { 871b6e66be2SVincenzo Maffione /* Stop polling from netmap and the eventfds, and deallocate 872b6e66be2SVincenzo Maffione * the poll context. */ 873*5faab778SVincenzo Maffione if (!busy_wait) { 874b6e66be2SVincenzo Maffione __set_current_state(TASK_RUNNING); 875*5faab778SVincenzo Maffione } 876b6e66be2SVincenzo Maffione for (i = 0; i < poll_ctx->next_entry; i++) { 877b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry *entry = 878b6e66be2SVincenzo Maffione poll_ctx->entries + i; 879b6e66be2SVincenzo Maffione 880b6e66be2SVincenzo Maffione if (entry->wqh) 881b6e66be2SVincenzo Maffione remove_wait_queue(entry->wqh, &entry->wait); 882b6e66be2SVincenzo Maffione /* We did not get a reference to the eventfds, but 883b6e66be2SVincenzo Maffione * don't do that on netmap file descriptors (since 884b6e66be2SVincenzo Maffione * a reference was not taken. */ 885b6e66be2SVincenzo Maffione if (entry->filp && entry->filp != priv->np_filp) 886b6e66be2SVincenzo Maffione fput(entry->filp); 887b6e66be2SVincenzo Maffione if (entry->irq_ctx) 888b6e66be2SVincenzo Maffione eventfd_ctx_put(entry->irq_ctx); 889b6e66be2SVincenzo Maffione if (entry->irq_filp) 890b6e66be2SVincenzo Maffione fput(entry->irq_filp); 891b6e66be2SVincenzo Maffione } 892b6e66be2SVincenzo Maffione nm_os_free(poll_ctx); 893b6e66be2SVincenzo Maffione poll_ctx = NULL; 894b6e66be2SVincenzo Maffione } 895b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 896b6e66be2SVincenzo Maffione 897f79ba6d7SVincenzo Maffione if (args) { 898f79ba6d7SVincenzo Maffione nm_os_free(args); 899f79ba6d7SVincenzo Maffione args = NULL; 900f79ba6d7SVincenzo Maffione } 901f79ba6d7SVincenzo Maffione 902b6e66be2SVincenzo Maffione /* Reset the kloop state. */ 903b6e66be2SVincenzo Maffione NMG_LOCK(); 904b6e66be2SVincenzo Maffione priv->np_kloop_state = 0; 905*5faab778SVincenzo Maffione if (na_could_sleep) { 906*5faab778SVincenzo Maffione na->na_flags |= NAF_BDG_MAYSLEEP; 907*5faab778SVincenzo Maffione } 908b6e66be2SVincenzo Maffione NMG_UNLOCK(); 909b6e66be2SVincenzo Maffione 910b6e66be2SVincenzo Maffione return err; 911b6e66be2SVincenzo Maffione } 912b6e66be2SVincenzo Maffione 913b6e66be2SVincenzo Maffione int 914b6e66be2SVincenzo Maffione netmap_sync_kloop_stop(struct netmap_priv_d *priv) 915b6e66be2SVincenzo Maffione { 916a56136a1SVincenzo Maffione struct netmap_adapter *na; 917b6e66be2SVincenzo Maffione bool running = true; 918b6e66be2SVincenzo Maffione int err = 0; 919b6e66be2SVincenzo Maffione 920a56136a1SVincenzo Maffione if (priv->np_nifp == NULL) { 921a56136a1SVincenzo Maffione return ENXIO; 922a56136a1SVincenzo Maffione } 923a56136a1SVincenzo Maffione mb(); /* make sure following reads are not from cache */ 924a56136a1SVincenzo Maffione 925a56136a1SVincenzo Maffione na = priv->np_na; 926a56136a1SVincenzo Maffione if (!nm_netmap_on(na)) { 927a56136a1SVincenzo Maffione return ENXIO; 928a56136a1SVincenzo Maffione } 929a56136a1SVincenzo Maffione 930a56136a1SVincenzo Maffione /* Set the kloop stopping flag. */ 931b6e66be2SVincenzo Maffione NMG_LOCK(); 932b6e66be2SVincenzo Maffione priv->np_kloop_state |= NM_SYNC_KLOOP_STOPPING; 933b6e66be2SVincenzo Maffione NMG_UNLOCK(); 934a56136a1SVincenzo Maffione 935a56136a1SVincenzo Maffione /* Send a notification to the kloop, in case it is blocked in 936a56136a1SVincenzo Maffione * schedule_timeout(). We can use either RX or TX, because the 937a56136a1SVincenzo Maffione * kloop is waiting on both. */ 938a56136a1SVincenzo Maffione nm_os_selwakeup(priv->np_si[NR_RX]); 939a56136a1SVincenzo Maffione 940a56136a1SVincenzo Maffione /* Wait for the kloop to actually terminate. */ 941b6e66be2SVincenzo Maffione while (running) { 942b6e66be2SVincenzo Maffione usleep_range(1000, 1500); 943b6e66be2SVincenzo Maffione NMG_LOCK(); 944b6e66be2SVincenzo Maffione running = (NM_ACCESS_ONCE(priv->np_kloop_state) 945b6e66be2SVincenzo Maffione & NM_SYNC_KLOOP_RUNNING); 946b6e66be2SVincenzo Maffione NMG_UNLOCK(); 947b6e66be2SVincenzo Maffione } 948b6e66be2SVincenzo Maffione 949b6e66be2SVincenzo Maffione return err; 950b6e66be2SVincenzo Maffione } 951b6e66be2SVincenzo Maffione 952b6e66be2SVincenzo Maffione #ifdef WITH_PTNETMAP 953b6e66be2SVincenzo Maffione /* 954b6e66be2SVincenzo Maffione * Guest ptnetmap txsync()/rxsync() routines, used in ptnet device drivers. 955b6e66be2SVincenzo Maffione * These routines are reused across the different operating systems supported 956b6e66be2SVincenzo Maffione * by netmap. 957b6e66be2SVincenzo Maffione */ 958b6e66be2SVincenzo Maffione 959b6e66be2SVincenzo Maffione /* 960b6e66be2SVincenzo Maffione * Reconcile host and guest views of the transmit ring. 961b6e66be2SVincenzo Maffione * 962b6e66be2SVincenzo Maffione * Guest user wants to transmit packets up to the one before ring->head, 963b6e66be2SVincenzo Maffione * and guest kernel knows tx_ring->hwcur is the first packet unsent 964b6e66be2SVincenzo Maffione * by the host kernel. 965b6e66be2SVincenzo Maffione * 966b6e66be2SVincenzo Maffione * We push out as many packets as possible, and possibly 967b6e66be2SVincenzo Maffione * reclaim buffers from previously completed transmission. 968b6e66be2SVincenzo Maffione * 969b6e66be2SVincenzo Maffione * Notifications from the host are enabled only if the user guest would 970b6e66be2SVincenzo Maffione * block (no space in the ring). 971b6e66be2SVincenzo Maffione */ 972b6e66be2SVincenzo Maffione bool 973b6e66be2SVincenzo Maffione netmap_pt_guest_txsync(struct nm_csb_atok *atok, struct nm_csb_ktoa *ktoa, 974b6e66be2SVincenzo Maffione struct netmap_kring *kring, int flags) 975b6e66be2SVincenzo Maffione { 976b6e66be2SVincenzo Maffione bool notify = false; 977b6e66be2SVincenzo Maffione 978b6e66be2SVincenzo Maffione /* Disable notifications */ 979b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 980b6e66be2SVincenzo Maffione 981b6e66be2SVincenzo Maffione /* 982*5faab778SVincenzo Maffione * First part: tell the host to process the new packets, 983*5faab778SVincenzo Maffione * updating the CSB. 984b6e66be2SVincenzo Maffione */ 985b6e66be2SVincenzo Maffione kring->nr_hwcur = ktoa->hwcur; 986f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_write(atok, kring->rcur, kring->rhead); 987b6e66be2SVincenzo Maffione 988b6e66be2SVincenzo Maffione /* Ask for a kick from a guest to the host if needed. */ 989*5faab778SVincenzo Maffione if (((kring->rhead != kring->nr_hwcur || nm_kr_wouldblock(kring)) 990b6e66be2SVincenzo Maffione && NM_ACCESS_ONCE(ktoa->kern_need_kick)) || 991b6e66be2SVincenzo Maffione (flags & NAF_FORCE_RECLAIM)) { 992b6e66be2SVincenzo Maffione atok->sync_flags = flags; 993b6e66be2SVincenzo Maffione notify = true; 994b6e66be2SVincenzo Maffione } 995b6e66be2SVincenzo Maffione 996b6e66be2SVincenzo Maffione /* 997b6e66be2SVincenzo Maffione * Second part: reclaim buffers for completed transmissions. 998b6e66be2SVincenzo Maffione */ 999*5faab778SVincenzo Maffione if (nm_kr_wouldblock(kring) || (flags & NAF_FORCE_RECLAIM)) { 1000f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 1001f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 1002b6e66be2SVincenzo Maffione } 1003b6e66be2SVincenzo Maffione 1004b6e66be2SVincenzo Maffione /* 1005b6e66be2SVincenzo Maffione * No more room in the ring for new transmissions. The user thread will 1006b6e66be2SVincenzo Maffione * go to sleep and we need to be notified by the host when more free 1007b6e66be2SVincenzo Maffione * space is available. 1008b6e66be2SVincenzo Maffione */ 1009*5faab778SVincenzo Maffione if (nm_kr_wouldblock(kring) && !(kring->nr_kflags & NKR_NOINTR)) { 1010b6e66be2SVincenzo Maffione /* Reenable notifications. */ 1011b6e66be2SVincenzo Maffione atok->appl_need_kick = 1; 1012f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 1013f79ba6d7SVincenzo Maffione nm_stld_barrier(); 1014f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 1015f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 1016b6e66be2SVincenzo Maffione /* If there is new free space, disable notifications */ 1017*5faab778SVincenzo Maffione if (unlikely(!nm_kr_wouldblock(kring))) { 1018b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 1019b6e66be2SVincenzo Maffione } 1020b6e66be2SVincenzo Maffione } 1021b6e66be2SVincenzo Maffione 1022b6e66be2SVincenzo Maffione nm_prdis(1, "%s CSB(head:%u cur:%u hwtail:%u) KRING(head:%u cur:%u tail:%u)", 1023b6e66be2SVincenzo Maffione kring->name, atok->head, atok->cur, ktoa->hwtail, 1024b6e66be2SVincenzo Maffione kring->rhead, kring->rcur, kring->nr_hwtail); 1025b6e66be2SVincenzo Maffione 1026b6e66be2SVincenzo Maffione return notify; 1027b6e66be2SVincenzo Maffione } 1028b6e66be2SVincenzo Maffione 1029b6e66be2SVincenzo Maffione /* 1030b6e66be2SVincenzo Maffione * Reconcile host and guest view of the receive ring. 1031b6e66be2SVincenzo Maffione * 1032b6e66be2SVincenzo Maffione * Update hwcur/hwtail from host (reading from CSB). 1033b6e66be2SVincenzo Maffione * 1034b6e66be2SVincenzo Maffione * If guest user has released buffers up to the one before ring->head, we 1035b6e66be2SVincenzo Maffione * also give them to the host. 1036b6e66be2SVincenzo Maffione * 1037b6e66be2SVincenzo Maffione * Notifications from the host are enabled only if the user guest would 1038b6e66be2SVincenzo Maffione * block (no more completed slots in the ring). 1039b6e66be2SVincenzo Maffione */ 1040b6e66be2SVincenzo Maffione bool 1041b6e66be2SVincenzo Maffione netmap_pt_guest_rxsync(struct nm_csb_atok *atok, struct nm_csb_ktoa *ktoa, 1042b6e66be2SVincenzo Maffione struct netmap_kring *kring, int flags) 1043b6e66be2SVincenzo Maffione { 1044b6e66be2SVincenzo Maffione bool notify = false; 1045b6e66be2SVincenzo Maffione 1046b6e66be2SVincenzo Maffione /* Disable notifications */ 1047b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 1048b6e66be2SVincenzo Maffione 1049b6e66be2SVincenzo Maffione /* 1050b6e66be2SVincenzo Maffione * First part: import newly received packets, by updating the kring 1051b6e66be2SVincenzo Maffione * hwtail to the hwtail known from the host (read from the CSB). 1052b6e66be2SVincenzo Maffione * This also updates the kring hwcur. 1053b6e66be2SVincenzo Maffione */ 1054f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, &kring->nr_hwcur); 1055b6e66be2SVincenzo Maffione kring->nr_kflags &= ~NKR_PENDINTR; 1056b6e66be2SVincenzo Maffione 1057b6e66be2SVincenzo Maffione /* 1058b6e66be2SVincenzo Maffione * Second part: tell the host about the slots that guest user has 1059b6e66be2SVincenzo Maffione * released, by updating cur and head in the CSB. 1060b6e66be2SVincenzo Maffione */ 1061b6e66be2SVincenzo Maffione if (kring->rhead != kring->nr_hwcur) { 1062f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_write(atok, kring->rcur, kring->rhead); 1063b6e66be2SVincenzo Maffione } 1064b6e66be2SVincenzo Maffione 1065b6e66be2SVincenzo Maffione /* 1066b6e66be2SVincenzo Maffione * No more completed RX slots. The user thread will go to sleep and 1067b6e66be2SVincenzo Maffione * we need to be notified by the host when more RX slots have been 1068b6e66be2SVincenzo Maffione * completed. 1069b6e66be2SVincenzo Maffione */ 1070*5faab778SVincenzo Maffione if (nm_kr_wouldblock(kring) && !(kring->nr_kflags & NKR_NOINTR)) { 1071b6e66be2SVincenzo Maffione /* Reenable notifications. */ 1072b6e66be2SVincenzo Maffione atok->appl_need_kick = 1; 1073f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 1074f79ba6d7SVincenzo Maffione nm_stld_barrier(); 1075f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 1076f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 1077b6e66be2SVincenzo Maffione /* If there are new slots, disable notifications. */ 1078*5faab778SVincenzo Maffione if (!nm_kr_wouldblock(kring)) { 1079b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 1080b6e66be2SVincenzo Maffione } 1081b6e66be2SVincenzo Maffione } 1082b6e66be2SVincenzo Maffione 1083*5faab778SVincenzo Maffione /* Ask for a kick from the guest to the host if needed. */ 1084*5faab778SVincenzo Maffione if ((kring->rhead != kring->nr_hwcur || nm_kr_wouldblock(kring)) 1085*5faab778SVincenzo Maffione && NM_ACCESS_ONCE(ktoa->kern_need_kick)) { 1086*5faab778SVincenzo Maffione atok->sync_flags = flags; 1087*5faab778SVincenzo Maffione notify = true; 1088*5faab778SVincenzo Maffione } 1089*5faab778SVincenzo Maffione 1090b6e66be2SVincenzo Maffione nm_prdis(1, "%s CSB(head:%u cur:%u hwtail:%u) KRING(head:%u cur:%u tail:%u)", 1091b6e66be2SVincenzo Maffione kring->name, atok->head, atok->cur, ktoa->hwtail, 1092b6e66be2SVincenzo Maffione kring->rhead, kring->rcur, kring->nr_hwtail); 1093b6e66be2SVincenzo Maffione 1094b6e66be2SVincenzo Maffione return notify; 1095b6e66be2SVincenzo Maffione } 1096b6e66be2SVincenzo Maffione 1097b6e66be2SVincenzo Maffione /* 1098b6e66be2SVincenzo Maffione * Callbacks for ptnet drivers: nm_krings_create, nm_krings_delete, nm_dtor. 1099b6e66be2SVincenzo Maffione */ 1100b6e66be2SVincenzo Maffione int 1101b6e66be2SVincenzo Maffione ptnet_nm_krings_create(struct netmap_adapter *na) 1102b6e66be2SVincenzo Maffione { 1103b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 1104b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; /* Upcast. */ 1105b6e66be2SVincenzo Maffione struct netmap_adapter *na_nm = &ptna->hwup.up; 1106b6e66be2SVincenzo Maffione struct netmap_adapter *na_dr = &ptna->dr.up; 1107b6e66be2SVincenzo Maffione int ret; 1108b6e66be2SVincenzo Maffione 1109b6e66be2SVincenzo Maffione if (ptna->backend_users) { 1110b6e66be2SVincenzo Maffione return 0; 1111b6e66be2SVincenzo Maffione } 1112b6e66be2SVincenzo Maffione 1113b6e66be2SVincenzo Maffione /* Create krings on the public netmap adapter. */ 1114b6e66be2SVincenzo Maffione ret = netmap_hw_krings_create(na_nm); 1115b6e66be2SVincenzo Maffione if (ret) { 1116b6e66be2SVincenzo Maffione return ret; 1117b6e66be2SVincenzo Maffione } 1118b6e66be2SVincenzo Maffione 1119b6e66be2SVincenzo Maffione /* Copy krings into the netmap adapter private to the driver. */ 1120b6e66be2SVincenzo Maffione na_dr->tx_rings = na_nm->tx_rings; 1121b6e66be2SVincenzo Maffione na_dr->rx_rings = na_nm->rx_rings; 1122b6e66be2SVincenzo Maffione 1123b6e66be2SVincenzo Maffione return 0; 1124b6e66be2SVincenzo Maffione } 1125b6e66be2SVincenzo Maffione 1126b6e66be2SVincenzo Maffione void 1127b6e66be2SVincenzo Maffione ptnet_nm_krings_delete(struct netmap_adapter *na) 1128b6e66be2SVincenzo Maffione { 1129b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 1130b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; /* Upcast. */ 1131b6e66be2SVincenzo Maffione struct netmap_adapter *na_nm = &ptna->hwup.up; 1132b6e66be2SVincenzo Maffione struct netmap_adapter *na_dr = &ptna->dr.up; 1133b6e66be2SVincenzo Maffione 1134b6e66be2SVincenzo Maffione if (ptna->backend_users) { 1135b6e66be2SVincenzo Maffione return; 1136b6e66be2SVincenzo Maffione } 1137b6e66be2SVincenzo Maffione 1138b6e66be2SVincenzo Maffione na_dr->tx_rings = NULL; 1139b6e66be2SVincenzo Maffione na_dr->rx_rings = NULL; 1140b6e66be2SVincenzo Maffione 1141b6e66be2SVincenzo Maffione netmap_hw_krings_delete(na_nm); 1142b6e66be2SVincenzo Maffione } 1143b6e66be2SVincenzo Maffione 1144b6e66be2SVincenzo Maffione void 1145b6e66be2SVincenzo Maffione ptnet_nm_dtor(struct netmap_adapter *na) 1146b6e66be2SVincenzo Maffione { 1147b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 1148b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; 1149b6e66be2SVincenzo Maffione 1150b6e66be2SVincenzo Maffione netmap_mem_put(ptna->dr.up.nm_mem); 1151b6e66be2SVincenzo Maffione memset(&ptna->dr, 0, sizeof(ptna->dr)); 1152b6e66be2SVincenzo Maffione netmap_mem_pt_guest_ifp_del(na->nm_mem, na->ifp); 1153b6e66be2SVincenzo Maffione } 1154b6e66be2SVincenzo Maffione 1155b6e66be2SVincenzo Maffione int 1156b6e66be2SVincenzo Maffione netmap_pt_guest_attach(struct netmap_adapter *arg, 1157b6e66be2SVincenzo Maffione unsigned int nifp_offset, unsigned int memid) 1158b6e66be2SVincenzo Maffione { 1159b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna; 1160b6e66be2SVincenzo Maffione struct ifnet *ifp = arg ? arg->ifp : NULL; 1161b6e66be2SVincenzo Maffione int error; 1162b6e66be2SVincenzo Maffione 1163b6e66be2SVincenzo Maffione /* get allocator */ 1164b6e66be2SVincenzo Maffione arg->nm_mem = netmap_mem_pt_guest_new(ifp, nifp_offset, memid); 1165b6e66be2SVincenzo Maffione if (arg->nm_mem == NULL) 1166b6e66be2SVincenzo Maffione return ENOMEM; 1167b6e66be2SVincenzo Maffione arg->na_flags |= NAF_MEM_OWNER; 1168b6e66be2SVincenzo Maffione error = netmap_attach_ext(arg, sizeof(struct netmap_pt_guest_adapter), 1); 1169b6e66be2SVincenzo Maffione if (error) 1170b6e66be2SVincenzo Maffione return error; 1171b6e66be2SVincenzo Maffione 1172b6e66be2SVincenzo Maffione /* get the netmap_pt_guest_adapter */ 1173b6e66be2SVincenzo Maffione ptna = (struct netmap_pt_guest_adapter *) NA(ifp); 1174b6e66be2SVincenzo Maffione 1175b6e66be2SVincenzo Maffione /* Initialize a separate pass-through netmap adapter that is going to 1176b6e66be2SVincenzo Maffione * be used by the ptnet driver only, and so never exposed to netmap 1177b6e66be2SVincenzo Maffione * applications. We only need a subset of the available fields. */ 1178b6e66be2SVincenzo Maffione memset(&ptna->dr, 0, sizeof(ptna->dr)); 1179b6e66be2SVincenzo Maffione ptna->dr.up.ifp = ifp; 1180b6e66be2SVincenzo Maffione ptna->dr.up.nm_mem = netmap_mem_get(ptna->hwup.up.nm_mem); 1181b6e66be2SVincenzo Maffione ptna->dr.up.nm_config = ptna->hwup.up.nm_config; 1182b6e66be2SVincenzo Maffione 1183b6e66be2SVincenzo Maffione ptna->backend_users = 0; 1184b6e66be2SVincenzo Maffione 1185b6e66be2SVincenzo Maffione return 0; 1186b6e66be2SVincenzo Maffione } 1187b6e66be2SVincenzo Maffione 1188b6e66be2SVincenzo Maffione #endif /* WITH_PTNETMAP */ 1189