Home
last modified time | relevance | path

Searched refs:new_chunk (Results 1 – 8 of 8) sorted by relevance

/linux/drivers/md/
H A Ddm-snap-persistent.c89 __le64 new_chunk; member
94 uint64_t new_chunk; member
416 result->new_chunk = le64_to_cpu(de->new_chunk); in read_exception()
426 de->new_chunk = cpu_to_le64(e->new_chunk); in write_exception()
435 de->new_chunk = 0; in clear_exception()
465 if (e.new_chunk == 0LL) { in insert_exceptions()
474 if (ps->next_free <= e.new_chunk) in insert_exceptions()
475 ps->next_free = e.new_chunk + 1; in insert_exceptions()
480 r = callback(callback_context, e.old_chunk, e.new_chunk); in insert_exceptions()
684 e->new_chunk = ps->next_free; in persistent_prepare_exception()
[all …]
H A Ddm-exception-store.h35 chunk_t new_chunk; member
150 return e->new_chunk >> DM_CHUNK_NUMBER_BITS; in dm_consecutive_chunk_count()
155 e->new_chunk += (1ULL << DM_CHUNK_NUMBER_BITS); in dm_consecutive_chunk_count_inc()
164 e->new_chunk -= (1ULL << DM_CHUNK_NUMBER_BITS); in dm_consecutive_chunk_count_dec()
H A Ddm-snap.c786 new_e->new_chunk == (dm_chunk_number(e->new_chunk) + in dm_insert_exception()
795 new_e->new_chunk == (dm_chunk_number(e->new_chunk) - 1)) { in dm_insert_exception()
798 e->new_chunk--; in dm_insert_exception()
839 e->new_chunk = new; in dm_add_exception()
965 e->new_chunk++; in __remove_single_exception_chunk()
1040 chunk_t old_chunk, new_chunk; in snapshot_merge_next_chunks() local
1058 &new_chunk); in snapshot_merge_next_chunks()
1071 new_chunk = new_chunk + 1 - linear_chunks; in snapshot_merge_next_chunks()
1084 src.sector = chunk_to_sector(s->store, new_chunk); in snapshot_merge_next_chunks()
1803 dest.sector = chunk_to_sector(s->store, pe->e.new_chunk); in start_copy()
[all …]
H A Ddm-snap-transient.c51 e->new_chunk = sector_to_chunk(store, tc->next_free); in transient_prepare_exception()
H A Draid5.c8755 int new_chunk = mddev->new_chunk_sectors; in raid5_check_reshape() local
8759 if (new_chunk > 0) { in raid5_check_reshape()
8760 if (!is_power_of_2(new_chunk)) in raid5_check_reshape()
8762 if (new_chunk < (PAGE_SIZE>>9)) in raid5_check_reshape()
8764 if (mddev->array_sectors & (new_chunk-1)) in raid5_check_reshape()
8777 if (new_chunk > 0) { in raid5_check_reshape()
8778 conf->chunk_sectors = new_chunk ; in raid5_check_reshape()
8779 mddev->chunk_sectors = new_chunk; in raid5_check_reshape()
8789 int new_chunk = mddev->new_chunk_sectors; in raid6_check_reshape() local
8793 if (new_chunk > 0) { in raid6_check_reshape()
[all …]
H A Dmd.c1508 mddev->new_chunk_sectors = sb->new_chunk >> 9; in super_90_validate()
1649 sb->new_chunk = mddev->new_chunk_sectors << 9; in super_90_sync()
2044 mddev->new_chunk_sectors = le32_to_cpu(sb->new_chunk); in super_1_validate()
2255 sb->new_chunk = cpu_to_le32(mddev->new_chunk_sectors); in super_1_sync()
/linux/include/uapi/linux/raid/
H A Dmd_p.h182 __u32 new_chunk; /* 17 new chunk size (bytes) */ member
260 __le32 new_chunk; /* new chunk size (512byte sectors) */ member
/linux/kernel/liveupdate/
H A Dkexec_handover.c342 static struct khoser_mem_chunk *new_chunk(struct khoser_mem_chunk *cur_chunk, in new_chunk() function
405 chunk = new_chunk(chunk, order); in kho_mem_serialize()
418 chunk = new_chunk(chunk, order); in kho_mem_serialize()