Lines Matching full:areas

125  * otherwise a segment may span two or more areas. It conflicts with free
145 * swiotlb_adjust_nareas() - adjust the number of areas and slots
146 * @nareas: Desired number of areas. Zero is treated as 1.
148 * Adjust the default number of areas in a memory pool.
168 * limit_nareas() - get the maximum number of areas for a given memory pool size
169 * @nareas: Desired number of areas.
172 * Limit the number of areas to the maximum possible number of areas in
175 * Return: Maximum possible number of areas.
283 spin_lock_init(&mem->areas[i].lock);
284 mem->areas[i].index = 0;
285 mem->areas[i].used = 0;
406 mem->areas = memblock_alloc(array_size(sizeof(struct io_tlb_area),
408 if (!mem->areas) {
409 pr_warn("%s: Failed to allocate mem->areas.\n", __func__);
498 area_order = get_order(array_size(sizeof(*mem->areas), nareas));
499 mem->areas = (struct io_tlb_area *)
501 if (!mem->areas)
519 free_pages((unsigned long)mem->areas, area_order);
545 area_order = get_order(array_size(sizeof(*mem->areas),
547 free_pages((unsigned long)mem->areas, area_order);
551 memblock_free_late(__pa(mem->areas),
552 array_size(sizeof(*mem->areas), mem->nareas));
673 * @nareas: Number of areas.
698 pool_size = sizeof(*pool) + array_size(sizeof(*pool->areas), nareas);
702 pool->areas = (void *)pool + sizeof(*pool);
948 * multiple areas.
1020 struct io_tlb_area *area = pool->areas + area_index;
1312 used += pool->areas[i].used;
1321 * areas.
1456 area = &mem->areas[aindex];
1825 pool->areas = kcalloc(nareas, sizeof(*pool->areas),
1827 if (!pool->areas) {