Lines Matching refs:map

41 static void	adj_cyl_offset(struct dk_map32 *map);
42 static int check_map(struct dk_map32 *map);
43 static void get_user_map(struct dk_map32 *map, int float_part);
44 static void get_user_map_efi(struct dk_gpt *map, int float_part);
60 struct dk_map32 *map = tmp_pinfo->pinfo_map; in p_modify() local
156 map[i].dkl_nblk = in p_modify()
158 map[i].dkl_cylno = in p_modify()
168 map[i].dkl_nblk = 0; in p_modify()
169 map[i].dkl_cylno = 0; in p_modify()
171 map[C_PARTITION].dkl_nblk = ncyl * spc(); in p_modify()
178 map[I_PARTITION].dkl_nblk = spc(); in p_modify()
179 map[I_PARTITION].dkl_cylno = 0; in p_modify()
181 map[J_PARTITION].dkl_nblk = 2 * spc(); in p_modify()
182 map[J_PARTITION].dkl_cylno = spc() / spc(); in p_modify()
232 map[free_hog].dkl_nblk = map[C_PARTITION].dkl_nblk; in p_modify()
234 map[free_hog].dkl_nblk -= map[I_PARTITION].dkl_nblk; in p_modify()
236 map[free_hog].dkl_nblk -= in p_modify()
237 map[J_PARTITION].dkl_nblk; in p_modify()
246 if (map[free_hog].dkl_nblk == 0) { in p_modify()
284 get_user_map(map, free_hog); in p_modify()
289 adj_cyl_offset(map); in p_modify()
304 cur_parts->pinfo_map[i].dkl_nblk = map[i].dkl_nblk; in p_modify()
305 cur_parts->pinfo_map[i].dkl_cylno = map[i].dkl_cylno; in p_modify()
308 map[i].dkl_cylno * nhead * nsect; in p_modify()
310 map[i].dkl_nblk; in p_modify()
334 adj_cyl_offset(struct dk_map32 *map) argument
350 if (i != C_PARTITION && map[i].dkl_nblk) {
351 map[i].dkl_cylno = cyloffset;
352 cyloffset += (map[i].dkl_nblk + (spc()-1))/spc();
353 } else if (map[i].dkl_nblk == 0) {
354 map[i].dkl_cylno = 0;
363 if (i != C_PARTITION && map[i].dkl_nblk) {
364 map[i].dkl_cylno = cyloffset;
365 cyloffset += (map[i].dkl_nblk + (spc()-1))/spc();
366 } else if (map[i].dkl_nblk == 0) {
367 map[i].dkl_cylno = 0;
377 check_map(struct dk_map32 *map) argument
387 cyloffset = map[0].dkl_cylno;
388 tot_blks = map[0].dkl_nblk;
396 if (map[i].dkl_cylno > (blkaddr32_t)ncyl-1) {
399 (PARTITION_BASE+i), map[i].dkl_cylno);
402 if (map[i].dkl_nblk >
403 (blkaddr32_t)(ncyl - map[i].dkl_cylno) * spc()) {
406 (PARTITION_BASE+i), map[i].dkl_nblk);
409 if (i != C_PARTITION && map[i].dkl_nblk) {
414 if (map[i].dkl_cylno < cyloffset) {
418 } else if (map[i].dkl_cylno > cyloffset) {
422 cyloffset += (map[i].dkl_nblk + (spc()-1))/spc();
423 tot_blks = map[i].dkl_nblk;
426 if (tot_blks > map[C_PARTITION].dkl_nblk) {
441 get_user_map(struct dk_map32 *map, int float_part) argument
459 ioparam.io_bounds.upper = map[i].dkl_nblk +
460 map[float_part].dkl_nblk;
461 deflt = map[i].dkl_nblk;
473 map[float_part].dkl_nblk -= (newsize - map[i].dkl_nblk);
474 map[i].dkl_nblk = newsize;
531 get_user_map_efi(struct dk_gpt *map, int float_part) argument
538 uint64_t start_lba = map->efi_first_u_lba;
541 reserved = efi_reserved_sectors(map);
542 for (i = 0; i < map->efi_nparts - 1; i++) {
548 ioparam.io_bounds.upper = map->efi_last_u_lba;
550 efi_deflt.end_sector = map->efi_parts[i].p_size;
555 map->efi_parts[i].p_tag = V_UNASSIGNED;
556 } else if ((i64 != 0) && (map->efi_parts[i].p_tag ==
558 map->efi_parts[i].p_tag = V_USR;
561 map->efi_parts[i].p_start = 0;
563 map->efi_parts[i].p_start = start_lba;
565 map->efi_parts[i].p_size = i64;
568 map->efi_parts[float_part].p_start = start_lba;
569 map->efi_parts[float_part].p_size = map->efi_last_u_lba + 1 -
571 map->efi_parts[float_part].p_tag = V_USR;
572 if (map->efi_parts[float_part].p_size == 0) {
573 map->efi_parts[float_part].p_size = 0;
574 map->efi_parts[float_part].p_start = 0;
575 map->efi_parts[float_part].p_tag = V_UNASSIGNED;
579 for (i = 0; i < map->efi_nparts; i++) {
580 if (map->efi_parts[i].p_tag == V_RESERVED) {
581 map->efi_parts[i].p_start = map->efi_last_u_lba -
583 map->efi_parts[i].p_size = reserved;