Lines Matching full:a64

222 		if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) {  in esas2r_build_sg_list_sge()
237 sgelen = (u8)((u8 *)sgc->sge.a64.curr in esas2r_build_sg_list_sge()
238 - (u8 *)sgc->sge.a64.last); in esas2r_build_sg_list_sge()
244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge()
247 sgc->sge.a64.curr = in esas2r_build_sg_list_sge()
252 sgc->sge.a64.limit = in esas2r_build_sg_list_sge()
257 sgc->sge.a64.last->length = cpu_to_le32( in esas2r_build_sg_list_sge()
259 sgc->sge.a64.last->address = in esas2r_build_sg_list_sge()
268 if (sgc->sge.a64.chain) { in esas2r_build_sg_list_sge()
269 sgc->sge.a64.chain->length |= in esas2r_build_sg_list_sge()
271 ((u8 *)(sgc->sge.a64. in esas2r_build_sg_list_sge()
280 sge.a64.last - in esas2r_build_sg_list_sge()
299 sgc->sge.a64.chain = sgc->sge.a64.last; in esas2r_build_sg_list_sge()
306 sgc->sge.a64.last = sgc->sge.a64.curr; in esas2r_build_sg_list_sge()
309 sgc->sge.a64.curr->length = cpu_to_le32(SGE_ADDR_64 | len); in esas2r_build_sg_list_sge()
310 sgc->sge.a64.curr->address = cpu_to_le32(addr); in esas2r_build_sg_list_sge()
311 sgc->sge.a64.curr++; in esas2r_build_sg_list_sge()
328 sgc->sge.a64.last->length |= cpu_to_le32(SGE_LAST); in esas2r_build_sg_list_sge()
334 if (sgc->sge.a64.chain) { in esas2r_build_sg_list_sge()
335 sgc->sge.a64.chain->length |= cpu_to_le32( in esas2r_build_sg_list_sge()
336 ((u8 *)(sgc->sge.a64.curr) - in esas2r_build_sg_list_sge()
346 ((u16)((u8 *)sgc->sge.a64.last - (u8 *)vrq) in esas2r_build_sg_list_sge()
614 (struct atto_physical_region_description *)sgc->sge.a64.curr; in esas2r_build_sg_list_prd()