Lines Matching refs:groupwidth
924 uint64_t groupwidth = vdc->vdc_groupwidth; in vdev_draid_logical_to_physical() local
933 uint64_t groupstart = (group * groupwidth) % ndisks; in vdev_draid_logical_to_physical()
934 ASSERT3U(groupstart + groupwidth, <=, ndisks + groupstart); in vdev_draid_logical_to_physical()
938 b_offset = b_offset % (rowheight_sectors * groupwidth); in vdev_draid_logical_to_physical()
939 ASSERT0(b_offset % groupwidth); in vdev_draid_logical_to_physical()
950 uint64_t row = (*perm * ((groupwidth * ngroups) / ndisks)) + in vdev_draid_logical_to_physical()
951 (((group % ngroups) * groupwidth) / ndisks); in vdev_draid_logical_to_physical()
954 (b_offset / groupwidth)) << ashift); in vdev_draid_logical_to_physical()
1000 uint64_t groupwidth = vdc->vdc_groupwidth; in vdev_draid_map_alloc_row() local
1001 uint64_t wrap = groupwidth; in vdev_draid_map_alloc_row()
1003 if (groupstart + groupwidth > ndisks) in vdev_draid_map_alloc_row()
1023 ASSERT3U(bc, <, groupwidth); in vdev_draid_map_alloc_row()
1030 raidz_row_t *rr = vdev_raidz_row_alloc(groupwidth, zio); in vdev_draid_map_alloc_row()
1042 for (uint64_t i = 0; i < groupwidth; i++) { in vdev_draid_map_alloc_row()
1064 rr->rr_nempty = roundup(tot, groupwidth) - tot; in vdev_draid_map_alloc_row()
1065 IMPLY(bc > 0, rr->rr_nempty == groupwidth - bc); in vdev_draid_map_alloc_row()