Home
last modified time | relevance | path

Searched refs:chunk (Results 1 – 25 of 300) sorted by relevance

12345678910>>...12

/linux/net/sctp/
H A Dinqueue.c42 static inline void sctp_inq_chunk_free(struct sctp_chunk *chunk) in sctp_inq_chunk_free() argument
44 if (chunk->head_skb) in sctp_inq_chunk_free()
45 chunk->skb = chunk->head_skb; in sctp_inq_chunk_free()
46 sctp_chunk_free(chunk); in sctp_inq_chunk_free()
52 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local
55 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free()
56 list_del_init(&chunk->list); in sctp_inq_free()
57 sctp_chunk_free(chunk); in sctp_inq_free()
72 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument
75 if (chunk->rcvr->dead) { in sctp_inq_push()
[all …]
H A Dchunk.c60 struct sctp_chunk *chunk; in sctp_datamsg_free() local
65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
66 sctp_chunk_free(chunk); in sctp_datamsg_free()
76 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local
83 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy()
86 sctp_chunk_put(chunk); in sctp_datamsg_destroy()
90 asoc = chunk->asoc; in sctp_datamsg_destroy()
92 sent = chunk->has_tsn ? SCTP_DATA_SENT : SCTP_DATA_UNSENT; in sctp_datamsg_destroy()
96 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy()
104 ev = sctp_ulpevent_make_send_failed_event(asoc, chunk, in sctp_datamsg_destroy()
[all …]
H A Doutput.c46 struct sctp_chunk *chunk);
48 struct sctp_chunk *chunk);
50 struct sctp_chunk *chunk);
52 struct sctp_chunk *chunk,
122 struct sctp_chunk *chunk = sctp_get_ecne_prepend(asoc); in sctp_packet_config() local
124 if (chunk) in sctp_packet_config()
125 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config()
163 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local
167 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
168 list_del_init(&chunk->list); in sctp_packet_free()
[all …]
H A Doutqueue.c210 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local
216 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
219 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
220 sctp_chunk_free(chunk); in __sctp_outq_teardown()
227 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
229 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
230 sctp_chunk_free(chunk); in __sctp_outq_teardown()
236 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
238 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
239 sctp_chunk_free(chunk); in __sctp_outq_teardown()
[all …]
H A Dsm_statefuns.c56 struct sctp_chunk *chunk,
59 struct sctp_chunk *chunk,
64 const struct sctp_chunk *chunk);
68 const struct sctp_chunk *chunk,
99 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk);
151 struct sctp_chunk *chunk);
174 static inline bool sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument
177 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid()
180 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid()
189 static inline bool sctp_err_chunk_valid(struct sctp_chunk *chunk) in sctp_err_chunk_valid() argument
[all …]
/linux/net/sunrpc/xprtrdma/
H A Dsvc_rdma_pcl.c20 struct svc_rdma_chunk *chunk; in pcl_free() local
22 chunk = pcl_first_chunk(pcl); in pcl_free()
23 list_del(&chunk->ch_list); in pcl_free()
24 kfree(chunk); in pcl_free()
30 struct svc_rdma_chunk *chunk; in pcl_alloc_chunk() local
32 chunk = kmalloc_flex(*chunk, ch_segments, segcount); in pcl_alloc_chunk()
33 if (!chunk) in pcl_alloc_chunk()
36 chunk->ch_position = position; in pcl_alloc_chunk()
37 chunk->ch_length = 0; in pcl_alloc_chunk()
38 chunk->ch_payload_length = 0; in pcl_alloc_chunk()
[all …]
/linux/mm/
H A Dpercpu-vm.c13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument
17 WARN_ON(chunk->immutable); in pcpu_chunk_page()
19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page()
54 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument
82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument
127 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument
131 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush()
132 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush()
153 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument
163 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages()
[all …]
H A Dpercpu.c28 * There is special consideration for the first chunk which must handle
30 * are not online yet. In short, the first chunk is structured like so:
45 * The allocator tries to allocate from the fullest chunk first. Each chunk
50 * of the bitmap. The reverse mapping from page to chunk is stored in
54 * Each bit represents a fragment of size PCPU_MIN_ALLOC_SIZE. The chunk
66 * setup the first chunk containing the kernel static percpu area
146 /* the address of the first chunk which starts with the kernel static area */
158 * The first chunk which always exists. Note that unlike other
165 * Optional reserved chunk
215 pcpu_addr_in_chunk(struct pcpu_chunk * chunk,void * addr) pcpu_addr_in_chunk() argument
242 pcpu_chunk_slot(const struct pcpu_chunk * chunk) pcpu_chunk_slot() argument
275 pcpu_chunk_addr(struct pcpu_chunk * chunk,unsigned int cpu,int page_idx) pcpu_chunk_addr() argument
286 pcpu_index_alloc_map(struct pcpu_chunk * chunk,int index) pcpu_index_alloc_map() argument
365 pcpu_next_md_free_region(struct pcpu_chunk * chunk,int * bit_off,int * bits) pcpu_next_md_free_region() argument
420 pcpu_next_fit_region(struct pcpu_chunk * chunk,int alloc_bits,int align,int * bit_off,int * bits) pcpu_next_fit_region() argument
477 pcpu_for_each_md_free_region(chunk,bit_off,bits) global() argument
483 pcpu_for_each_fit_region(chunk,alloc_bits,align,bit_off,bits) global() argument
526 __pcpu_chunk_move(struct pcpu_chunk * chunk,int slot,bool move_front) __pcpu_chunk_move() argument
537 pcpu_chunk_move(struct pcpu_chunk * chunk,int slot) pcpu_chunk_move() argument
555 pcpu_chunk_relocate(struct pcpu_chunk * chunk,int oslot) pcpu_chunk_relocate() argument
567 pcpu_isolate_chunk(struct pcpu_chunk * chunk) pcpu_isolate_chunk() argument
578 pcpu_reintegrate_chunk(struct pcpu_chunk * chunk) pcpu_reintegrate_chunk() argument
598 pcpu_update_empty_pages(struct pcpu_chunk * chunk,int nr) pcpu_update_empty_pages() argument
712 pcpu_block_update_scan(struct pcpu_chunk * chunk,int bit_off,int bits) pcpu_block_update_scan() argument
745 pcpu_chunk_refresh_hint(struct pcpu_chunk * chunk,bool full_scan) pcpu_chunk_refresh_hint() argument
774 pcpu_block_refresh_hint(struct pcpu_chunk * chunk,int index) pcpu_block_refresh_hint() argument
808 pcpu_block_update_hint_alloc(struct pcpu_chunk * chunk,int bit_off,int bits) pcpu_block_update_hint_alloc() argument
963 pcpu_block_update_hint_free(struct pcpu_chunk * chunk,int bit_off,int bits) pcpu_block_update_hint_free() argument
1073 pcpu_is_populated(struct pcpu_chunk * chunk,int bit_off,int bits,int * next_off) pcpu_is_populated() argument
1110 pcpu_find_block_fit(struct pcpu_chunk * chunk,int alloc_bits,size_t align,bool pop_only) pcpu_find_block_fit() argument
1216 pcpu_alloc_area(struct pcpu_chunk * chunk,int alloc_bits,size_t align,int start) pcpu_alloc_area() argument
1276 pcpu_free_area(struct pcpu_chunk * chunk,int off) pcpu_free_area() argument
1319 pcpu_init_md_blocks(struct pcpu_chunk * chunk) pcpu_init_md_blocks() argument
1348 struct pcpu_chunk *chunk; pcpu_alloc_first_chunk() local
1427 struct pcpu_chunk *chunk; pcpu_alloc_chunk() local
1484 pcpu_free_chunk(struct pcpu_chunk * chunk) pcpu_free_chunk() argument
1507 pcpu_chunk_populated(struct pcpu_chunk * chunk,int page_start,int page_end) pcpu_chunk_populated() argument
1531 pcpu_chunk_depopulated(struct pcpu_chunk * chunk,int page_start,int page_end) pcpu_chunk_depopulated() argument
1630 pcpu_memcg_post_alloc_hook(struct obj_cgroup * objcg,struct pcpu_chunk * chunk,int off,size_t size) pcpu_memcg_post_alloc_hook() argument
1649 pcpu_memcg_free_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_memcg_free_hook() argument
1679 pcpu_memcg_post_alloc_hook(struct obj_cgroup * objcg,struct pcpu_chunk * chunk,int off,size_t size) pcpu_memcg_post_alloc_hook() argument
1684 pcpu_memcg_free_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_memcg_free_hook() argument
1690 pcpu_alloc_tag_alloc_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_alloc_tag_alloc_hook() argument
1699 pcpu_alloc_tag_free_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_alloc_tag_free_hook() argument
1705 pcpu_alloc_tag_alloc_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_alloc_tag_alloc_hook() argument
1710 pcpu_alloc_tag_free_hook(struct pcpu_chunk * chunk,int off,size_t size) pcpu_alloc_tag_free_hook() argument
1738 struct pcpu_chunk *chunk, *next; pcpu_alloc_noprof() local
1949 struct pcpu_chunk *chunk, *next; pcpu_balance_free() local
2003 struct pcpu_chunk *chunk; pcpu_balance_populated() local
2094 struct pcpu_chunk *chunk; pcpu_reclaim_populated() local
2229 struct pcpu_chunk *chunk; free_percpu() local
[all...]
H A Dpercpu-km.c35 static void pcpu_post_unmap_tlb_flush(struct pcpu_chunk *chunk, in pcpu_post_unmap_tlb_flush() argument
41 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument
47 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument
56 struct pcpu_chunk *chunk; in pcpu_create_chunk() local
61 chunk = pcpu_alloc_chunk(gfp); in pcpu_create_chunk()
62 if (!chunk) in pcpu_create_chunk()
67 pcpu_free_chunk(chunk); in pcpu_create_chunk()
72 pcpu_set_page_chunk(pages + i, chunk); in pcpu_create_chunk()
74 chunk->data = pages; in pcpu_create_chunk()
75 chunk->base_addr = page_address(pages); in pcpu_create_chunk()
[all …]
/linux/drivers/s390/cio/
H A Ditcw.c184 void *chunk; in itcw_init() local
196 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init()
197 if (IS_ERR(chunk)) in itcw_init()
198 return chunk; in itcw_init()
199 itcw = chunk; in itcw_init()
212 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
213 if (IS_ERR(chunk)) in itcw_init()
214 return chunk; in itcw_init()
215 itcw->tcw = chunk; in itcw_init()
220 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
[all …]
/linux/kernel/trace/
H A Dpid_list.c15 union lower_chunk *chunk; in get_lower_chunk() local
22 chunk = pid_list->lower_list; in get_lower_chunk()
23 pid_list->lower_list = chunk->next; in get_lower_chunk()
26 chunk->next = NULL; in get_lower_chunk()
34 return chunk; in get_lower_chunk()
39 union upper_chunk *chunk; in get_upper_chunk() local
46 chunk = pid_list->upper_list; in get_upper_chunk()
47 pid_list->upper_list = chunk->next; in get_upper_chunk()
50 chunk->next = NULL; in get_upper_chunk()
58 return chunk; in get_upper_chunk()
[all …]
/linux/fs/xfs/
H A Dxfs_zone_gc.c586 struct xfs_gc_bio *chunk = in xfs_zone_gc_end_io() local
588 struct xfs_zone_gc_data *data = chunk->data; in xfs_zone_gc_end_io()
590 WRITE_ONCE(chunk->state, XFS_GC_BIO_DONE); in xfs_zone_gc_end_io()
641 struct xfs_gc_bio *chunk) in xfs_zone_gc_add_data() argument
643 struct xfs_zone_gc_data *data = chunk->data; in xfs_zone_gc_add_data()
644 unsigned int len = chunk->len; in xfs_zone_gc_add_data()
651 bio_add_folio_nofail(&chunk->bio, in xfs_zone_gc_add_data()
670 struct xfs_gc_bio *chunk; in xfs_zone_gc_start_chunk() local
694 chunk = container_of(bio, struct xfs_gc_bio, bio); in xfs_zone_gc_start_chunk()
695 chunk->ip = ip; in xfs_zone_gc_start_chunk()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dicm.c55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument
59 if (chunk->nsg > 0) in mlx4_free_icm_pages()
60 dma_unmap_sg(&dev->persist->pdev->dev, chunk->sg, chunk->npages, in mlx4_free_icm_pages()
63 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages()
64 __free_pages(sg_page(&chunk->sg[i]), in mlx4_free_icm_pages()
65 get_order(chunk->sg[i].length)); in mlx4_free_icm_pages()
68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument
72 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent()
74 chunk->buf[i].size, in mlx4_free_icm_coherent()
75 chunk->buf[i].addr, in mlx4_free_icm_coherent()
[all …]
H A Dicm.h74 struct mlx4_icm_chunk *chunk; member
100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
108 return !iter->chunk; in mlx4_icm_last()
113 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next()
114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
115 iter->chunk = NULL; in mlx4_icm_next()
119 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next()
127 if (iter->chunk->coherent) in mlx4_icm_addr()
128 return iter->chunk->buf[iter->page_idx].dma_addr; in mlx4_icm_addr()
130 return sg_dma_address(&iter->chunk->sg[iter->page_idx]); in mlx4_icm_addr()
[all …]
/linux/kernel/
H A Daudit_tree.c42 struct audit_chunk *chunk; member
130 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument
134 for (i = 0; i < chunk->count; i++) { in free_chunk()
135 if (chunk->owners[i].owner) in free_chunk()
136 put_tree(chunk->owners[i].owner); in free_chunk()
138 kfree(chunk); in free_chunk()
141 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument
143 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk()
144 free_chunk(chunk); in audit_put_chunk()
149 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local
[all …]
/linux/lib/
H A Dgenalloc.c40 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument
42 return chunk->end_addr - chunk->start_addr + 1; in chunk_size()
187 struct gen_pool_chunk *chunk; in gen_pool_add_owner() local
192 chunk = vzalloc_node(nbytes, nid); in gen_pool_add_owner()
193 if (unlikely(chunk == NULL)) in gen_pool_add_owner()
196 chunk->phys_addr = phys; in gen_pool_add_owner()
197 chunk->start_addr = virt; in gen_pool_add_owner()
198 chunk->end_addr = virt + size - 1; in gen_pool_add_owner()
199 chunk->owner = owner; in gen_pool_add_owner()
200 atomic_long_set(&chunk->avail, size); in gen_pool_add_owner()
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_ring_mux.c79 struct amdgpu_mux_chunk *chunk; in amdgpu_mux_resubmit_chunks() local
103 list_for_each_entry(chunk, &e->list, entry) { in amdgpu_mux_resubmit_chunks()
104 if (chunk->sync_seq > last_seq && chunk->sync_seq <= seq) { in amdgpu_mux_resubmit_chunks()
106 chunk->sync_seq, in amdgpu_mux_resubmit_chunks()
108 if (chunk->sync_seq == in amdgpu_mux_resubmit_chunks()
110 if (chunk->cntl_offset <= e->ring->buf_mask) in amdgpu_mux_resubmit_chunks()
112 chunk->cntl_offset); in amdgpu_mux_resubmit_chunks()
113 if (chunk->ce_offset <= e->ring->buf_mask) in amdgpu_mux_resubmit_chunks()
114 amdgpu_ring_patch_ce(e->ring, chunk->ce_offset); in amdgpu_mux_resubmit_chunks()
115 if (chunk->de_offset <= e->ring->buf_mask) in amdgpu_mux_resubmit_chunks()
[all …]
/linux/drivers/infiniband/hw/irdma/
H A Dpble.c18 struct irdma_chunk *chunk; in irdma_destroy_pble_prm() local
22 chunk = (struct irdma_chunk *) pinfo->clist.next; in irdma_destroy_pble_prm()
23 list_del(&chunk->list); in irdma_destroy_pble_prm()
24 if (chunk->type == PBLE_SD_PAGED) in irdma_destroy_pble_prm()
25 irdma_pble_free_paged_mem(chunk); in irdma_destroy_pble_prm()
26 bitmap_free(chunk->bitmapbuf); in irdma_destroy_pble_prm()
27 kfree(chunk->chunkmem.va); in irdma_destroy_pble_prm()
90 struct irdma_chunk *chunk = info->chunk; in add_sd_direct() local
103 chunk->type = PBLE_SD_CONTIGOUS; in add_sd_direct()
107 chunk->size = info->pages << HMC_PAGED_BP_SHIFT; in add_sd_direct()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_icm_pool.c80 u64 mlx5dr_icm_pool_get_chunk_mr_addr(struct mlx5dr_icm_chunk *chunk) in mlx5dr_icm_pool_get_chunk_mr_addr() argument
82 u32 offset = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_mr_addr()
84 return (u64)offset * chunk->seg; in mlx5dr_icm_pool_get_chunk_mr_addr()
87 u32 mlx5dr_icm_pool_get_chunk_rkey(struct mlx5dr_icm_chunk *chunk) in mlx5dr_icm_pool_get_chunk_rkey() argument
89 return chunk->buddy_mem->icm_mr->mkey; in mlx5dr_icm_pool_get_chunk_rkey()
92 u64 mlx5dr_icm_pool_get_chunk_icm_addr(struct mlx5dr_icm_chunk *chunk) in mlx5dr_icm_pool_get_chunk_icm_addr() argument
94 u32 size = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_icm_addr()
96 return (u64)chunk->buddy_mem->icm_mr->icm_start_addr + size * chunk->seg; in mlx5dr_icm_pool_get_chunk_icm_addr()
99 u32 mlx5dr_icm_pool_get_chunk_byte_size(struct mlx5dr_icm_chunk *chunk) in mlx5dr_icm_pool_get_chunk_byte_size() argument
101 return mlx5dr_icm_pool_chunk_size_to_byte(chunk->size, in mlx5dr_icm_pool_get_chunk_byte_size()
[all …]
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_memfree.c64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument
68 if (chunk->nsg > 0) in mthca_free_icm_pages()
69 dma_unmap_sg(&dev->pdev->dev, chunk->mem, chunk->npages, in mthca_free_icm_pages()
72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages()
73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages()
74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages()
77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument
81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent()
82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent()
83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent()
[all …]
/linux/include/net/sctp/
H A Dsm.h77 /* Prototypes for chunk state functions. */
168 /* Prototypes for chunk-building functions. */
173 const struct sctp_chunk *chunk,
176 const struct sctp_chunk *chunk);
178 const struct sctp_chunk *chunk);
181 const struct sctp_chunk *chunk);
194 const struct sctp_chunk *chunk);
196 const struct sctp_chunk *chunk);
199 const struct sctp_chunk *chunk);
200 int sctp_init_cause(struct sctp_chunk *chunk, __be1
334 sctp_data_size(struct sctp_chunk * chunk) sctp_data_size() argument
375 sctp_vtag_verify(const struct sctp_chunk * chunk,const struct sctp_association * asoc) sctp_vtag_verify() argument
395 sctp_vtag_verify_either(const struct sctp_chunk * chunk,const struct sctp_association * asoc) sctp_vtag_verify_either() argument
[all...]
/linux/kernel/liveupdate/
H A Dkexec_handover.c363 struct khoser_mem_chunk *chunk __free(free_page) = NULL; in new_chunk()
365 chunk = (void *)get_zeroed_page(GFP_KERNEL); in new_chunk()
366 if (!chunk) in new_chunk()
369 if (WARN_ON(kho_scratch_overlap(virt_to_phys(chunk), PAGE_SIZE))) in new_chunk()
372 chunk->hdr.order = order; in new_chunk()
374 KHOSER_STORE_PTR(cur_chunk->hdr.next, chunk); in new_chunk()
375 return no_free_ptr(chunk); in new_chunk()
380 struct khoser_mem_chunk *chunk = first_chunk; in kho_mem_ser_free() local
382 while (chunk) { in kho_mem_ser_free()
383 struct khoser_mem_chunk *tmp = chunk; in kho_mem_ser_free()
[all …]
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_heap.c127 struct panthor_heap_chunk *chunk) in panthor_free_heap_chunk() argument
130 list_del(&chunk->node); in panthor_free_heap_chunk()
136 panthor_kernel_bo_destroy(chunk->bo); in panthor_free_heap_chunk()
137 kfree(chunk); in panthor_free_heap_chunk()
144 struct panthor_heap_chunk *chunk; in panthor_alloc_heap_chunk() local
148 chunk = kmalloc_obj(*chunk); in panthor_alloc_heap_chunk()
149 if (!chunk) in panthor_alloc_heap_chunk()
152 chunk->bo = panthor_kernel_bo_create(pool->ptdev, pool->vm, heap->chunk_size, in panthor_alloc_heap_chunk()
157 if (IS_ERR(chunk->bo)) { in panthor_alloc_heap_chunk()
158 ret = PTR_ERR(chunk->bo); in panthor_alloc_heap_chunk()
[all …]
/linux/drivers/gpu/drm/panel/
H A Dpanel-samsung-s6e63m0-dsi.c44 int chunk; in s6e63m0_dsi_dcs_write() local
54 chunk = remain; in s6e63m0_dsi_dcs_write()
57 if (chunk > S6E63M0_DSI_MAX_CHUNK) in s6e63m0_dsi_dcs_write()
58 chunk = S6E63M0_DSI_MAX_CHUNK; in s6e63m0_dsi_dcs_write()
59 ret = mipi_dsi_dcs_write(dsi, cmd, seqp, chunk); in s6e63m0_dsi_dcs_write()
64 cmdwritten += chunk; in s6e63m0_dsi_dcs_write()
65 seqp += chunk; in s6e63m0_dsi_dcs_write()
68 chunk = remain - cmdwritten; in s6e63m0_dsi_dcs_write()
69 if (chunk > S6E63M0_DSI_MAX_CHUNK) in s6e63m0_dsi_dcs_write()
70 chunk = S6E63M0_DSI_MAX_CHUNK; in s6e63m0_dsi_dcs_write()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c104 struct nouveau_dmem_chunk *chunk = nouveau_page_to_chunk(page); in page_to_drm() local
106 return chunk->drm; in page_to_drm()
111 struct nouveau_dmem_chunk *chunk = nouveau_page_to_chunk(page); in nouveau_dmem_page_addr() local
113 chunk->pagemap.range.start; in nouveau_dmem_page_addr()
115 return chunk->bo->offset + off; in nouveau_dmem_page_addr()
121 struct nouveau_dmem_chunk *chunk = nouveau_page_to_chunk(page); in nouveau_dmem_folio_free() local
122 struct nouveau_dmem *dmem = chunk->drm->dmem; in nouveau_dmem_folio_free()
133 WARN_ON(!chunk->callocated); in nouveau_dmem_folio_free()
134 chunk->callocated--; in nouveau_dmem_folio_free()
299 struct nouveau_dmem_chunk *chunk; in nouveau_dmem_chunk_alloc() local
[all …]

12345678910>>...12