Lines Matching defs:pg_vec

524 	h.raw = rb->pg_vec[pg_vec_pos].buffer +
635 struct pgv *pg_vec,
644 p1->pkbdq = pg_vec;
645 pbd = (struct tpacket_block_desc *)pg_vec[0].buffer;
646 p1->pkblk_start = pg_vec[0].buffer;
2513 if (likely(po->tx_ring.pg_vec)) {
2738 /* packet_sendmsg() check on tx_ring.pg_vec was lockless,
2741 if (unlikely(!po->tx_ring.pg_vec)) {
3102 /* Reading tx_ring.pg_vec without holding pg_vec_lock is racy.
3105 if (data_race(po->tx_ring.pg_vec))
3149 if (po->rx_ring.pg_vec) {
3154 if (po->tx_ring.pg_vec) {
3871 if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) {
3891 if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) {
3910 if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) {
3964 if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) {
4028 if (!po->rx_ring.pg_vec && !po->tx_ring.pg_vec)
4296 if (po->rx_ring.pg_vec) {
4304 if (po->tx_ring.pg_vec) {
4342 static void free_pg_vec(struct pgv *pg_vec, unsigned int order,
4348 if (likely(pg_vec[i].buffer)) {
4349 if (is_vmalloc_addr(pg_vec[i].buffer))
4350 vfree(pg_vec[i].buffer);
4352 free_pages((unsigned long)pg_vec[i].buffer,
4354 pg_vec[i].buffer = NULL;
4357 kfree(pg_vec);
4388 struct pgv *pg_vec;
4391 pg_vec = kcalloc(block_nr, sizeof(struct pgv), GFP_KERNEL | __GFP_NOWARN);
4392 if (unlikely(!pg_vec))
4396 pg_vec[i].buffer = alloc_one_pg_vec_page(order);
4397 if (unlikely(!pg_vec[i].buffer))
4402 return pg_vec;
4405 free_pg_vec(pg_vec, order, block_nr);
4406 pg_vec = NULL;
4413 struct pgv *pg_vec = NULL;
4440 if (unlikely(rb->pg_vec))
4481 pg_vec = alloc_pg_vec(req, order);
4482 if (unlikely(!pg_vec))
4488 init_prb_bdqc(po, rb, pg_vec, req_u);
4535 swap(rb->pg_vec, pg_vec);
4547 po->prot_hook.func = (po->rx_ring.pg_vec) ?
4562 if (pg_vec && (po->tp_version > TPACKET_V2)) {
4569 if (pg_vec) {
4571 free_pg_vec(pg_vec, order, req->tp_block_nr);
4595 if (rb->pg_vec) {
4611 if (rb->pg_vec == NULL)
4616 void *kaddr = rb->pg_vec[i].buffer;