Lines Matching defs:pfrag
258 struct page_frag *pfrag,
264 if (skb_frag_page(frag) == pfrag->page &&
265 skb_frag_off(frag) + skb_frag_size(frag) == pfrag->offset) {
269 skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
272 get_page(pfrag->page);
275 pfrag->offset += size;
316 struct page_frag *pfrag,
328 if (unlikely(pfrag->size - pfrag->offset < prot->tag_size) &&
329 !skb_page_frag_refill(prot->tag_size, pfrag, sk->sk_allocation)) {
332 pfrag = &dummy_tag_frag;
334 tls_append_frag(record, pfrag, prot->tag_size);
343 struct page_frag *pfrag,
354 skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
357 get_page(pfrag->page);
358 pfrag->offset += prepend_size;
368 struct page_frag *pfrag,
374 if (unlikely(!skb_page_frag_refill(prepend_size, pfrag,
381 ret = tls_create_new_record(offload_ctx, pfrag, prepend_size);
385 if (pfrag->size > pfrag->offset)
389 if (!sk_page_frag_refill(sk, pfrag))
430 struct page_frag *pfrag;
459 pfrag = sk_page_frag(sk);
467 rc = tls_do_allocation(sk, ctx, pfrag, prot->prepend_size);
519 copy = min_t(size_t, copy, pfrag->size - pfrag->offset);
521 rc = tls_device_copy_data(page_address(pfrag->page) +
522 pfrag->offset, copy,
526 tls_append_frag(record, pfrag, copy);
544 pfrag, record_type);