| /linux/include/net/ |
| H A D | fq_impl.h | 36 idx = flow - fq->flows; in __fq_adjust_removal() 152 flow = &fq->flows[idx]; in fq_flow_classify() 160 tin->flows++; in fq_flow_classify() 173 struct fq_flow *cur = &fq->flows[i]; in fq_find_fattest_flow() 361 fq->flows = kvcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL); in fq_init() 362 if (!fq->flows) in fq_init() 367 kvfree(fq->flows); in fq_init() 368 fq->flows = NULL; in fq_init() 373 fq_flow_init(&fq->flows[i]); in fq_init() 384 fq_flow_reset(fq, &fq->flows[i], free_func); in fq_reset() [all …]
|
| H A D | fq.h | 57 u32 flows; member 69 struct fq_flow *flows; member
|
| /linux/samples/bpf/ |
| H A D | do_hbm_test.sh | 18 echo " [-f=<#flows>|--flows=<#flows>] [-h] [-i=<id>|--id=<id >]" 34 echo " -f or --flows number of concurrent flows (default=1)" 38 echo " -l do not limit flows using loopback" 78 flows=1 150 -f=*|--flows=*) 151 flows="${i#*=}" 278 while [ $flow_cnt -le $flows ] ; d [all...] |
| /linux/net/sched/ |
| H A D | sch_fq_codel.c | 29 * Packets are classified (internal classifier or external) on flows. 30 * This is a Stochastic model (as we use a hash, several flows 34 * so that new flows have priority on old ones. 53 struct fq_codel_flow *flows; /* Flows table [flows_cnt] */ member 55 u32 flows_cnt; /* number of flows */ 66 struct list_head new_flows; /* list of new flows */ 67 struct list_head old_flows; /* list of old flows */ 148 * This might sound expensive, but with 1024 flows, we scan in fq_codel_drop() 164 flow = &q->flows[idx]; in fq_codel_drop() 205 flow = &q->flows[id in fq_codel_enqueue() [all...] |
| H A D | sch_fq_pie.c | 58 struct fq_pie_flow *flows; member 152 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue() 307 if (q->flows) { in fq_pie_change() 403 &q->flows[q->flows_cursor].vars, in fq_pie_timer() 404 q->flows[q->flows_cursor].backlog); in fq_pie_timer() 451 q->flows = kvcalloc(q->flows_cnt, sizeof(struct fq_pie_flow), in fq_pie_init() 453 if (!q->flows) { in fq_pie_init() 458 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_init() 543 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_reset() 561 kvfree(q->flows); in fq_pie_destroy()
|
| H A D | sch_cake.c | 29 * flows from each other. This prevents a burst on one flow from increasing 34 * Codel and Blue AQM algorithms. This serves flows fairly, and signals 151 struct cake_flow flows[CAKE_QUEUES]; member 268 * obtain the best features of each. Codel is excellent on flows which 270 * unresponsive flows. 787 q->flows[reduced_hash].set)) { in cake_hash() 805 if (!q->flows[outer_hash + k].set) { in cake_hash() 820 if (!q->flows[outer_hash + k].set) { in cake_hash() 835 if (q->flows[outer_hash + k].set == CAKE_SET_BULK) { in cake_hash() 836 cake_dec_srchost_bulk_flow_count(q, &q->flows[outer_has in cake_hash() [all...] |
| H A D | sch_fq.c | 13 * They are also part of one Round Robin 'queues' (new or old flows) 27 * dequeue() : serves flows in Round Robin 138 struct rb_root delayed; /* for rate limited flows */ 143 u32 flows; member 248 /* limit number of collected flows per round */ 295 q->flows -= fcnt; in fq_gc() 324 * scheduled in the future (ie no flows are eligible) in fq_fastpath_check() 327 if (q->flows != q->inactive_flows + q->throttled_flows) in fq_fastpath_check() 336 /* Ordering invariants fall apart if some delayed flows in fq_fastpath_check() 458 q->flows in fq_classify() [all...] |
| /linux/drivers/crypto/allwinner/sun8i-ss/ |
| H A D | sun8i-ss-prng.c | 134 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_prng_generate() 135 ss->flows[flow].status = 0; in sun8i_ss_prng_generate() 141 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_prng_generate() 143 if (ss->flows[flow].status == 0) { in sun8i_ss_prng_generate()
|
| H A D | sun8i-ss-hash.c | 295 ss->flows[flow].stat_req++; in sun8i_ss_run_hash_task() 328 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_hash_task() 329 ss->flows[flow].status = 0; in sun8i_ss_run_hash_task() 334 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_hash_task() 336 if (ss->flows[flow].status == 0) { in sun8i_ss_run_hash_task() 412 engine = ss->flows[e].engine; in sun8i_ss_hash_digest() 511 result = ss->flows[rctx->flow].result; in sun8i_ss_hash_run() 512 pad = ss->flows[rctx->flow].pad; in sun8i_ss_hash_run()
|
| H A D | sun8i-ss.h | 163 struct sun8i_ss_flow *flows; member
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | tid_rdma.c | 757 u32 generation = rcd->flows[flow_idx].generation; in kern_setup_hw_flow() 775 rcd->flows[flow_idx].generation = in kern_clear_hw_flow() 776 kern_flow_generation_next(rcd->flows[flow_idx].generation); in kern_clear_hw_flow() 804 rcd->flows[fs->index].generation = fs->generation; in hfi1_kern_setup_hw_flow() 853 rcd->flows[i].generation = mask_generation(get_random_u32()); in hfi1_kern_init_ctxt_generations() 1465 struct tid_rdma_flow *flow = &req->flows[req->setup_head]; in hfi1_kern_exp_rcv_setup() 1556 struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; in hfi1_kern_exp_rcv_clear() 1612 kfree(req->flows); in hfi1_kern_exp_rcv_free_flows() 1613 req->flows = NULL; in hfi1_kern_exp_rcv_free_flows() 1634 struct tid_rdma_flow *flows; in hfi1_kern_exp_rcv_alloc_flows() local [all …]
|
| /linux/Documentation/networking/ |
| H A D | nf_flowtable.rst | 33 specifies what flows are placed into the flowtable. Hence, packets follow the 34 classic IP forwarding path unless the user explicitly instruct flows to use this 111 You can identify offloaded flows through the [OFFLOAD] tag when listing your 130 instead the real device is sufficient for the flowtable to track your flows. 198 There is a workqueue that adds the flows to the hardware. Note that a few 202 You can identify hardware offloaded flows through the [HW_OFFLOAD] tag when
|
| H A D | scaling.rst | 31 of logical flows. Packets for each flow are steered to a separate receive 50 applications that monitor TCP/IP flows (IDS, firewalls, ...etc) and need 252 to the same CPU is CPU load imbalance if flows vary in packet rate. 258 Flow Limit is an optional RPS feature that prioritizes small flows 259 during CPU contention by dropping packets from large flows slightly 260 ahead of those from small flows. It is active only when an RPS or RFS 266 new packet is dropped. Packets from other flows are still only 270 even large flows maintain connectivity. 288 identification of large flows and fewer false positives. The default 325 flows to the CPUs where those flows are being processed. The flow hash [all …]
|
| H A D | openvswitch.rst | 16 table" that userspace populates with "flows" that map from keys based 104 A wildcarded flow can represent a group of exact match flows. Each '1' bit 108 by reduce the number of new flows need to be processed by the user space program. 120 two possible approaches: reactively install flows as they miss the kernel 130 The behavior when using overlapping wildcarded flows is undefined. It is the 133 performs best-effort detection of overlapping wildcarded flows and may reject 146 future operations. The kernel is not required to index flows by the original
|
| H A D | pktgen.rst | 97 flows: 0 flowlen: 0 112 flows: 0 286 pgset "flows 1" 391 flows
|
| /linux/Documentation/userspace-api/media/mediactl/ |
| H A D | media-controller-model.rst | 26 by an entity flows from the entity's output to one or more entity 31 pads, either on the same entity or on different entities. Data flows
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_rep.h | 183 struct list_head flows; member 208 struct list_head flows; member
|
| /linux/Documentation/admin-guide/pm/ |
| H A D | system-wide.rst | 11 suspend-flows
|
| H A D | suspend-flows.rst | 25 The kernel code flows associated with the suspend and resume transitions for 27 significant differences between the :ref:`suspend-to-idle <s2idle>` code flows 28 and the code flows related to the :ref:`suspend-to-RAM <s2ram>` and 35 available. Apart from that, the suspend and resume code flows for these sleep
|
| /linux/Documentation/admin-guide/blockdev/drbd/ |
| H A D | figures.rst | 5 Data flows that Relate some functions, and write packets
|
| /linux/net/mctp/test/ |
| H A D | route-test.c | 987 struct mctp_flow *flows[2]; in mctp_test_packet_flow() 1011 flows[0] = skb_ext_find(tx_skbs[0], SKB_EXT_MCTP); in mctp_test_fragment_flow() 1012 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, flows[0]); in mctp_test_fragment_flow() local 1013 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, flows[0]->key); in mctp_test_fragment_flow() 1014 KUNIT_ASSERT_PTR_EQ(test, flows[0]->key->sk, sock->sk); in mctp_test_fragment_flow() 1016 flows[1] = skb_ext_find(tx_skbs[1], SKB_EXT_MCTP); in mctp_test_fragment_flow() 1017 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, flows[1]); in mctp_test_fragment_flow() 1018 KUNIT_ASSERT_PTR_EQ(test, flows[1]->key, flows[0]->key); in mctp_test_fragment_flow()
|
| /linux/drivers/cache/ |
| H A D | Kconfig | 40 These drivers implement cache management for flows where it is necessary
|
| /linux/net/dccp/ccids/ |
| H A D | Kconfig | |
| /linux/arch/arm/boot/dts/ti/keystone/ |
| H A D | keystone-k2g-netcp.dtsi | 85 <0x4013000 0x400>; /* 32 Rx flows */
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-platform-kim | 9 The device name flows down to architecture specific board
|