xref: /freebsd/sys/dev/irdma/irdma_pble.c (revision 01fbb86991d524288a785b544c29d9c5ea1b61fb)
1 /*-
2  * SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
3  *
4  * Copyright (c) 2015 - 2023 Intel Corporation
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenFabrics.org BSD license below:
11  *
12  *   Redistribution and use in source and binary forms, with or
13  *   without modification, are permitted provided that the following
14  *   conditions are met:
15  *
16  *    - Redistributions of source code must retain the above
17  *	copyright notice, this list of conditions and the following
18  *	disclaimer.
19  *
20  *    - Redistributions in binary form must reproduce the above
21  *	copyright notice, this list of conditions and the following
22  *	disclaimer in the documentation and/or other materials
23  *	provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  */
34 
35 #include "osdep.h"
36 #include "irdma_hmc.h"
37 #include "irdma_defs.h"
38 #include "irdma_type.h"
39 #include "irdma_protos.h"
40 #include "irdma_pble.h"
41 
42 static int add_pble_prm(struct irdma_hmc_pble_rsrc *pble_rsrc);
43 
44 /**
45  * irdma_destroy_pble_prm - destroy prm during module unload
46  * @pble_rsrc: pble resources
47  */
48 void
irdma_destroy_pble_prm(struct irdma_hmc_pble_rsrc * pble_rsrc)49 irdma_destroy_pble_prm(struct irdma_hmc_pble_rsrc *pble_rsrc)
50 {
51 	struct irdma_chunk *chunk;
52 	struct irdma_pble_prm *pinfo = &pble_rsrc->pinfo;
53 
54 	while (!list_empty(&pinfo->clist)) {
55 		chunk = (struct irdma_chunk *)(&pinfo->clist)->next;
56 		list_del(&chunk->list);
57 		if (chunk->type == PBLE_SD_PAGED)
58 			irdma_pble_free_paged_mem(chunk);
59 		bitmap_free(chunk->bitmapbuf);
60 		kfree(chunk->chunkmem.va);
61 	}
62 	spin_lock_destroy(&pinfo->prm_lock);
63 	mutex_destroy(&pble_rsrc->pble_mutex_lock);
64 }
65 
66 /**
67  * irdma_hmc_init_pble - Initialize pble resources during module load
68  * @dev: irdma_sc_dev struct
69  * @pble_rsrc: pble resources
70  */
71 int
irdma_hmc_init_pble(struct irdma_sc_dev * dev,struct irdma_hmc_pble_rsrc * pble_rsrc)72 irdma_hmc_init_pble(struct irdma_sc_dev *dev,
73 		    struct irdma_hmc_pble_rsrc *pble_rsrc)
74 {
75 	struct irdma_hmc_info *hmc_info;
76 	u32 fpm_idx = 0;
77 	int status = 0;
78 
79 	hmc_info = dev->hmc_info;
80 	pble_rsrc->dev = dev;
81 	pble_rsrc->fpm_base_addr = hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].base;
82 	/* Start pble' on 4k boundary */
83 	if (pble_rsrc->fpm_base_addr & 0xfff)
84 		fpm_idx = (4096 - (pble_rsrc->fpm_base_addr & 0xfff)) >> 3;
85 	pble_rsrc->unallocated_pble =
86 	    hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt - fpm_idx;
87 	pble_rsrc->next_fpm_addr = pble_rsrc->fpm_base_addr + (fpm_idx << 3);
88 	pble_rsrc->pinfo.pble_shift = PBLE_SHIFT;
89 
90 	mutex_init(&pble_rsrc->pble_mutex_lock);
91 
92 	spin_lock_init(&pble_rsrc->pinfo.prm_lock);
93 	INIT_LIST_HEAD(&pble_rsrc->pinfo.clist);
94 	if (add_pble_prm(pble_rsrc)) {
95 		irdma_destroy_pble_prm(pble_rsrc);
96 		status = -ENOMEM;
97 	}
98 
99 	return status;
100 }
101 
102 /**
103  * get_sd_pd_idx -  Returns sd index, pd index and rel_pd_idx from fpm address
104  * @pble_rsrc: structure containing fpm address
105  * @idx: where to return indexes
106  */
107 static void
get_sd_pd_idx(struct irdma_hmc_pble_rsrc * pble_rsrc,struct sd_pd_idx * idx)108 get_sd_pd_idx(struct irdma_hmc_pble_rsrc *pble_rsrc,
109 	      struct sd_pd_idx *idx)
110 {
111 	idx->sd_idx = (u32)pble_rsrc->next_fpm_addr / IRDMA_HMC_DIRECT_BP_SIZE;
112 	idx->pd_idx = (u32)(pble_rsrc->next_fpm_addr / IRDMA_HMC_PAGED_BP_SIZE);
113 	idx->rel_pd_idx = (idx->pd_idx % IRDMA_HMC_PD_CNT_IN_SD);
114 }
115 
116 /**
117  * add_sd_direct - add sd direct for pble
118  * @pble_rsrc: pble resource ptr
119  * @info: page info for sd
120  */
121 static int
add_sd_direct(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_add_page_info * info)122 add_sd_direct(struct irdma_hmc_pble_rsrc *pble_rsrc,
123 	      struct irdma_add_page_info *info)
124 {
125 	struct irdma_sc_dev *dev = pble_rsrc->dev;
126 	int ret_code = 0;
127 	struct sd_pd_idx *idx = &info->idx;
128 	struct irdma_chunk *chunk = info->chunk;
129 	struct irdma_hmc_info *hmc_info = info->hmc_info;
130 	struct irdma_hmc_sd_entry *sd_entry = info->sd_entry;
131 	u32 offset = 0;
132 
133 	if (!sd_entry->valid) {
134 		ret_code = irdma_add_sd_table_entry(dev->hw, hmc_info,
135 						    info->idx.sd_idx,
136 						    IRDMA_SD_TYPE_DIRECT,
137 						    IRDMA_HMC_DIRECT_BP_SIZE);
138 		if (ret_code)
139 			return ret_code;
140 
141 		chunk->type = PBLE_SD_CONTIGOUS;
142 	}
143 
144 	offset = idx->rel_pd_idx << HMC_PAGED_BP_SHIFT;
145 	chunk->size = info->pages << HMC_PAGED_BP_SHIFT;
146 	chunk->vaddr = (u8 *)sd_entry->u.bp.addr.va + offset;
147 	chunk->fpm_addr = pble_rsrc->next_fpm_addr;
148 	irdma_debug(dev, IRDMA_DEBUG_PBLE,
149 		    "chunk_size[%ld] = 0x%lx vaddr=0x%p fpm_addr = %lx\n",
150 		    chunk->size, chunk->size, chunk->vaddr, chunk->fpm_addr);
151 
152 	return 0;
153 }
154 
155 /**
156  * fpm_to_idx - given fpm address, get pble index
157  * @pble_rsrc: pble resource management
158  * @addr: fpm address for index
159  */
fpm_to_idx(struct irdma_hmc_pble_rsrc * pble_rsrc,u64 addr)160 static u32 fpm_to_idx(struct irdma_hmc_pble_rsrc *pble_rsrc, u64 addr){
161 	u64 idx;
162 
163 	idx = (addr - (pble_rsrc->fpm_base_addr)) >> 3;
164 
165 	return (u32)idx;
166 }
167 
168 /**
169  * add_bp_pages - add backing pages for sd
170  * @pble_rsrc: pble resource management
171  * @info: page info for sd
172  */
173 static int
add_bp_pages(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_add_page_info * info)174 add_bp_pages(struct irdma_hmc_pble_rsrc *pble_rsrc,
175 	     struct irdma_add_page_info *info)
176 {
177 	struct irdma_sc_dev *dev = pble_rsrc->dev;
178 	u8 *addr;
179 	struct irdma_dma_mem mem;
180 	struct irdma_hmc_pd_entry *pd_entry;
181 	struct irdma_hmc_sd_entry *sd_entry = info->sd_entry;
182 	struct irdma_hmc_info *hmc_info = info->hmc_info;
183 	struct irdma_chunk *chunk = info->chunk;
184 	int status = 0;
185 	u32 rel_pd_idx = info->idx.rel_pd_idx;
186 	u32 pd_idx = info->idx.pd_idx;
187 	u32 i;
188 
189 	if (irdma_pble_get_paged_mem(chunk, info->pages))
190 		return -ENOMEM;
191 
192 	status = irdma_add_sd_table_entry(dev->hw, hmc_info, info->idx.sd_idx,
193 					  IRDMA_SD_TYPE_PAGED,
194 					  IRDMA_HMC_DIRECT_BP_SIZE);
195 	if (status)
196 		goto error;
197 
198 	addr = chunk->vaddr;
199 	for (i = 0; i < info->pages; i++) {
200 		mem.pa = (u64)chunk->dmainfo.dmaaddrs[i];
201 		mem.size = 4096;
202 		mem.va = addr;
203 		pd_entry = &sd_entry->u.pd_table.pd_entry[rel_pd_idx++];
204 		if (!pd_entry->valid) {
205 			status = irdma_add_pd_table_entry(dev, hmc_info,
206 							  pd_idx++, &mem);
207 			if (status)
208 				goto error;
209 
210 			addr += 4096;
211 		}
212 	}
213 
214 	chunk->fpm_addr = pble_rsrc->next_fpm_addr;
215 	return 0;
216 
217 error:
218 	irdma_pble_free_paged_mem(chunk);
219 
220 	return status;
221 }
222 
223 /**
224  * irdma_get_type - add a sd entry type for sd
225  * @dev: irdma_sc_dev struct
226  * @idx: index of sd
227  * @pages: pages in the sd
228  */
229 static enum irdma_sd_entry_type
irdma_get_type(struct irdma_sc_dev * dev,struct sd_pd_idx * idx,u32 pages)230 irdma_get_type(struct irdma_sc_dev *dev,
231 	       struct sd_pd_idx *idx, u32 pages)
232 {
233 	enum irdma_sd_entry_type sd_entry_type;
234 
235 	sd_entry_type = !idx->rel_pd_idx && pages == IRDMA_HMC_PD_CNT_IN_SD ?
236 	    IRDMA_SD_TYPE_DIRECT : IRDMA_SD_TYPE_PAGED;
237 	return sd_entry_type;
238 }
239 
240 /**
241  * add_pble_prm - add a sd entry for pble resoure
242  * @pble_rsrc: pble resource management
243  */
244 static int
add_pble_prm(struct irdma_hmc_pble_rsrc * pble_rsrc)245 add_pble_prm(struct irdma_hmc_pble_rsrc *pble_rsrc)
246 {
247 	struct irdma_sc_dev *dev = pble_rsrc->dev;
248 	struct irdma_hmc_sd_entry *sd_entry;
249 	struct irdma_hmc_info *hmc_info;
250 	struct irdma_chunk *chunk;
251 	struct irdma_add_page_info info;
252 	struct sd_pd_idx *idx = &info.idx;
253 	int ret_code = 0;
254 	enum irdma_sd_entry_type sd_entry_type;
255 	u64 sd_reg_val = 0;
256 	struct irdma_virt_mem chunkmem;
257 	u32 pages;
258 
259 	if (pble_rsrc->unallocated_pble < PBLE_PER_PAGE)
260 		return -ENOMEM;
261 
262 	if (pble_rsrc->next_fpm_addr & 0xfff)
263 		return -EINVAL;
264 
265 	chunkmem.size = sizeof(*chunk);
266 	chunkmem.va = kzalloc(chunkmem.size, GFP_KERNEL);
267 	if (!chunkmem.va)
268 		return -ENOMEM;
269 
270 	chunk = chunkmem.va;
271 	chunk->chunkmem = chunkmem;
272 	hmc_info = dev->hmc_info;
273 	chunk->dev = dev;
274 	chunk->fpm_addr = pble_rsrc->next_fpm_addr;
275 	get_sd_pd_idx(pble_rsrc, idx);
276 	sd_entry = &hmc_info->sd_table.sd_entry[idx->sd_idx];
277 	pages = (idx->rel_pd_idx) ? (IRDMA_HMC_PD_CNT_IN_SD - idx->rel_pd_idx) :
278 	    IRDMA_HMC_PD_CNT_IN_SD;
279 	pages = min(pages, pble_rsrc->unallocated_pble >> PBLE_512_SHIFT);
280 	info.chunk = chunk;
281 	info.hmc_info = hmc_info;
282 	info.pages = pages;
283 	info.sd_entry = sd_entry;
284 	if (!sd_entry->valid)
285 		sd_entry_type = irdma_get_type(dev, idx, pages);
286 	else
287 		sd_entry_type = sd_entry->entry_type;
288 
289 	irdma_debug(dev, IRDMA_DEBUG_PBLE,
290 		    "pages = %d, unallocated_pble[%d] current_fpm_addr = %lx\n",
291 		    pages, pble_rsrc->unallocated_pble,
292 		    pble_rsrc->next_fpm_addr);
293 	irdma_debug(dev, IRDMA_DEBUG_PBLE, "sd_entry_type = %d\n",
294 		    sd_entry_type);
295 	if (sd_entry_type == IRDMA_SD_TYPE_DIRECT)
296 		ret_code = add_sd_direct(pble_rsrc, &info);
297 
298 	if (ret_code)
299 		sd_entry_type = IRDMA_SD_TYPE_PAGED;
300 	else
301 		pble_rsrc->stats_direct_sds++;
302 
303 	if (sd_entry_type == IRDMA_SD_TYPE_PAGED) {
304 		ret_code = add_bp_pages(pble_rsrc, &info);
305 		if (ret_code)
306 			goto err_bp_pages;
307 		else
308 			pble_rsrc->stats_paged_sds++;
309 	}
310 
311 	ret_code = irdma_prm_add_pble_mem(&pble_rsrc->pinfo, chunk);
312 	if (ret_code)
313 		goto err_bp_pages;
314 
315 	pble_rsrc->next_fpm_addr += chunk->size;
316 	irdma_debug(dev, IRDMA_DEBUG_PBLE,
317 		    "next_fpm_addr = %lx chunk_size[%lu] = 0x%lx\n",
318 		    pble_rsrc->next_fpm_addr, chunk->size, chunk->size);
319 	pble_rsrc->unallocated_pble -= (u32)(chunk->size >> 3);
320 	sd_reg_val = (sd_entry_type == IRDMA_SD_TYPE_PAGED) ?
321 	    sd_entry->u.pd_table.pd_page_addr.pa :
322 	    sd_entry->u.bp.addr.pa;
323 	if (!sd_entry->valid) {
324 		ret_code = irdma_hmc_sd_one(dev, hmc_info->hmc_fn_id, sd_reg_val,
325 					    idx->sd_idx, sd_entry->entry_type, true);
326 		if (ret_code)
327 			goto error;
328 	}
329 
330 	sd_entry->valid = true;
331 	list_add(&chunk->list, &pble_rsrc->pinfo.clist);
332 	return 0;
333 
334 error:
335 	bitmap_free(chunk->bitmapbuf);
336 err_bp_pages:
337 	kfree(chunk->chunkmem.va);
338 
339 	return ret_code;
340 }
341 
342 /**
343  * free_lvl2 - fee level 2 pble
344  * @pble_rsrc: pble resource management
345  * @palloc: level 2 pble allocation
346  */
347 static void
free_lvl2(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc)348 free_lvl2(struct irdma_hmc_pble_rsrc *pble_rsrc,
349 	  struct irdma_pble_alloc *palloc)
350 {
351 	u32 i;
352 	struct irdma_pble_level2 *lvl2 = &palloc->level2;
353 	struct irdma_pble_info *root = &lvl2->root;
354 	struct irdma_pble_info *leaf = lvl2->leaf;
355 
356 	for (i = 0; i < lvl2->leaf_cnt; i++, leaf++) {
357 		if (leaf->addr)
358 			irdma_prm_return_pbles(&pble_rsrc->pinfo,
359 					       &leaf->chunkinfo);
360 		else
361 			break;
362 	}
363 
364 	if (root->addr)
365 		irdma_prm_return_pbles(&pble_rsrc->pinfo, &root->chunkinfo);
366 
367 	kfree(lvl2->leafmem.va);
368 	lvl2->leaf = NULL;
369 }
370 
371 /**
372  * get_lvl2_pble - get level 2 pble resource
373  * @pble_rsrc: pble resource management
374  * @palloc: level 2 pble allocation
375  */
376 static int
get_lvl2_pble(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc)377 get_lvl2_pble(struct irdma_hmc_pble_rsrc *pble_rsrc,
378 	      struct irdma_pble_alloc *palloc)
379 {
380 	u32 lf4k, lflast, total, i;
381 	u32 pblcnt = PBLE_PER_PAGE;
382 	u64 *addr;
383 	struct irdma_pble_level2 *lvl2 = &palloc->level2;
384 	struct irdma_pble_info *root = &lvl2->root;
385 	struct irdma_pble_info *leaf;
386 	int ret_code;
387 	u64 fpm_addr;
388 
389 	/* number of full 512 (4K) leafs) */
390 	lf4k = palloc->total_cnt >> 9;
391 	lflast = palloc->total_cnt % PBLE_PER_PAGE;
392 	total = (lflast == 0) ? lf4k : lf4k + 1;
393 	lvl2->leaf_cnt = total;
394 
395 	lvl2->leafmem.size = (sizeof(*leaf) * total);
396 	lvl2->leafmem.va = kzalloc(lvl2->leafmem.size, GFP_KERNEL);
397 	if (!lvl2->leafmem.va)
398 		return -ENOMEM;
399 
400 	lvl2->leaf = lvl2->leafmem.va;
401 	leaf = lvl2->leaf;
402 	ret_code = irdma_prm_get_pbles(&pble_rsrc->pinfo, &root->chunkinfo,
403 				       total << 3, &root->addr, &fpm_addr);
404 	if (ret_code) {
405 		kfree(lvl2->leafmem.va);
406 		lvl2->leaf = NULL;
407 		return -ENOMEM;
408 	}
409 
410 	root->idx = fpm_to_idx(pble_rsrc, fpm_addr);
411 	root->cnt = total;
412 	addr = root->addr;
413 	for (i = 0; i < total; i++, leaf++) {
414 		pblcnt = (lflast && ((i + 1) == total)) ?
415 		    lflast : PBLE_PER_PAGE;
416 		ret_code = irdma_prm_get_pbles(&pble_rsrc->pinfo,
417 					       &leaf->chunkinfo, pblcnt << 3,
418 					       &leaf->addr, &fpm_addr);
419 		if (ret_code)
420 			goto error;
421 
422 		leaf->idx = fpm_to_idx(pble_rsrc, fpm_addr);
423 
424 		leaf->cnt = pblcnt;
425 		*addr = (u64)leaf->idx;
426 		addr++;
427 	}
428 
429 	palloc->level = PBLE_LEVEL_2;
430 	pble_rsrc->stats_lvl2++;
431 	return 0;
432 
433 error:
434 	free_lvl2(pble_rsrc, palloc);
435 
436 	return -ENOMEM;
437 }
438 
439 /**
440  * get_lvl1_pble - get level 1 pble resource
441  * @pble_rsrc: pble resource management
442  * @palloc: level 1 pble allocation
443  */
444 static int
get_lvl1_pble(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc)445 get_lvl1_pble(struct irdma_hmc_pble_rsrc *pble_rsrc,
446 	      struct irdma_pble_alloc *palloc)
447 {
448 	int ret_code;
449 	u64 fpm_addr;
450 	struct irdma_pble_info *lvl1 = &palloc->level1;
451 
452 	ret_code = irdma_prm_get_pbles(&pble_rsrc->pinfo, &lvl1->chunkinfo,
453 				       palloc->total_cnt << 3, &lvl1->addr,
454 				       &fpm_addr);
455 	if (ret_code)
456 		return -ENOMEM;
457 
458 	palloc->level = PBLE_LEVEL_1;
459 	lvl1->idx = fpm_to_idx(pble_rsrc, fpm_addr);
460 	lvl1->cnt = palloc->total_cnt;
461 	pble_rsrc->stats_lvl1++;
462 
463 	return 0;
464 }
465 
466 /**
467  * get_lvl1_lvl2_pble - calls get_lvl1 and get_lvl2 pble routine
468  * @pble_rsrc: pble resources
469  * @palloc: contains all inforamtion regarding pble (idx + pble addr)
470  * @lvl: Bitmask for requested pble level
471  */
472 static int
get_lvl1_lvl2_pble(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc,u8 lvl)473 get_lvl1_lvl2_pble(struct irdma_hmc_pble_rsrc *pble_rsrc,
474 		   struct irdma_pble_alloc *palloc, u8 lvl)
475 {
476 	int status = 0;
477 
478 	status = get_lvl1_pble(pble_rsrc, palloc);
479 	if (!status || lvl == PBLE_LEVEL_1 || palloc->total_cnt <= PBLE_PER_PAGE)
480 		return status;
481 
482 	status = get_lvl2_pble(pble_rsrc, palloc);
483 
484 	return status;
485 }
486 
487 /**
488  * irdma_get_pble - allocate pbles from the prm
489  * @pble_rsrc: pble resources
490  * @palloc: contains all inforamtion regarding pble (idx + pble addr)
491  * @pble_cnt: #of pbles requested
492  * @lvl: requested pble level mask
493  */
494 int
irdma_get_pble(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc,u32 pble_cnt,u8 lvl)495 irdma_get_pble(struct irdma_hmc_pble_rsrc *pble_rsrc,
496 	       struct irdma_pble_alloc *palloc, u32 pble_cnt,
497 	       u8 lvl)
498 {
499 	int status = 0;
500 	int max_sds = 0;
501 	int i;
502 
503 	palloc->total_cnt = pble_cnt;
504 	palloc->level = PBLE_LEVEL_0;
505 
506 	mutex_lock(&pble_rsrc->pble_mutex_lock);
507 
508 	/*
509 	 * check first to see if we can get pble's without acquiring additional sd's
510 	 */
511 	status = get_lvl1_lvl2_pble(pble_rsrc, palloc, lvl);
512 	if (!status)
513 		goto exit;
514 
515 	max_sds = (palloc->total_cnt >> 18) + 1;
516 	for (i = 0; i < max_sds; i++) {
517 		status = add_pble_prm(pble_rsrc);
518 		if (status)
519 			break;
520 
521 		status = get_lvl1_lvl2_pble(pble_rsrc, palloc, lvl);
522 		/* if level1_only, only go through it once */
523 		if (!status || lvl == PBLE_LEVEL_1)
524 			break;
525 	}
526 
527 exit:
528 	if (!status) {
529 		pble_rsrc->allocdpbles += pble_cnt;
530 		pble_rsrc->stats_alloc_ok++;
531 	} else {
532 		pble_rsrc->stats_alloc_fail++;
533 	}
534 	mutex_unlock(&pble_rsrc->pble_mutex_lock);
535 
536 	return status;
537 }
538 
539 /**
540  * irdma_free_pble - put pbles back into prm
541  * @pble_rsrc: pble resources
542  * @palloc: contains all information regarding pble resource being freed
543  */
544 void
irdma_free_pble(struct irdma_hmc_pble_rsrc * pble_rsrc,struct irdma_pble_alloc * palloc)545 irdma_free_pble(struct irdma_hmc_pble_rsrc *pble_rsrc,
546 		struct irdma_pble_alloc *palloc)
547 {
548 	pble_rsrc->freedpbles += palloc->total_cnt;
549 
550 	if (palloc->level == PBLE_LEVEL_2)
551 		free_lvl2(pble_rsrc, palloc);
552 	else
553 		irdma_prm_return_pbles(&pble_rsrc->pinfo,
554 				       &palloc->level1.chunkinfo);
555 	pble_rsrc->stats_alloc_freed++;
556 }
557