Home
last modified time | relevance | path

Searched refs:new_entry (Results 1 – 25 of 35) sorted by relevance

12

/linux/fs/ext4/
H A Dblock_validity.c73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local
88 new_entry = kmem_cache_alloc(ext4_system_zone_cachep, in add_system_zone()
90 if (!new_entry) in add_system_zone()
92 new_entry->start_blk = start_blk; in add_system_zone()
93 new_entry->count = count; in add_system_zone()
94 new_entry->ino = ino; in add_system_zone()
95 new_node = &new_entry->node; in add_system_zone()
104 if (can_merge(entry, new_entry)) { in add_system_zone()
105 new_entry->start_blk = entry->start_blk; in add_system_zone()
106 new_entry->count += entry->count; in add_system_zone()
[all …]
H A Dmballoc.c6440 struct ext4_free_data *new_entry) in ext4_mb_free_metadata() argument
6444 ext4_grpblk_t clusters = new_entry->efd_count; in ext4_mb_free_metadata()
6457 new_node = &new_entry->efd_node; in ext4_mb_free_metadata()
6458 cluster = new_entry->efd_start_cluster; in ext4_mb_free_metadata()
6481 kmem_cache_free(ext4_free_data_cachep, new_entry); in ext4_mb_free_metadata()
6491 if (ext4_freed_extents_can_be_merged(new_entry, entry)) { in ext4_mb_free_metadata()
6493 entry->efd_count += new_entry->efd_count; in ext4_mb_free_metadata()
6494 kmem_cache_free(ext4_free_data_cachep, new_entry); in ext4_mb_free_metadata()
6498 if (ext4_freed_extents_can_be_merged(entry, new_entry)) { in ext4_mb_free_metadata()
6499 entry->efd_count += new_entry->efd_count; in ext4_mb_free_metadata()
[all …]
/linux/drivers/iommu/generic_pt/
H A Dpt_fmt_defaults.h224 u64 new_entry; in pt_set_sw_bit_release() local
227 new_entry = old_entry | pt_sw_bit(bitnr); in pt_set_sw_bit_release()
228 } while (!try_cmpxchg64_release(entryp, &old_entry, new_entry)); in pt_set_sw_bit_release()
229 pts->entry = new_entry; in pt_set_sw_bit_release()
236 u32 new_entry; in pt_set_sw_bit_release() local
239 new_entry = old_entry | pt_sw_bit(bitnr); in pt_set_sw_bit_release()
240 } while (!try_cmpxchg_release(entryp, &old_entry, new_entry)); in pt_set_sw_bit_release()
241 pts->entry = new_entry; in pt_set_sw_bit_release()
/linux/mm/
H A Dshmem_quota.c170 struct quota_id *new_entry, *entry; in shmem_acquire_dquot() local
193 new_entry = kzalloc_obj(struct quota_id, GFP_NOFS); in shmem_acquire_dquot()
194 if (!new_entry) { in shmem_acquire_dquot()
199 new_entry->id = id; in shmem_acquire_dquot()
201 new_entry->bhardlimit = sbinfo->qlimits.usrquota_bhardlimit; in shmem_acquire_dquot()
202 new_entry->ihardlimit = sbinfo->qlimits.usrquota_ihardlimit; in shmem_acquire_dquot()
204 new_entry->bhardlimit = sbinfo->qlimits.grpquota_bhardlimit; in shmem_acquire_dquot()
205 new_entry->ihardlimit = sbinfo->qlimits.grpquota_ihardlimit; in shmem_acquire_dquot()
208 new_node = &new_entry->node; in shmem_acquire_dquot()
211 entry = new_entry; in shmem_acquire_dquot()
H A Dmremap.c597 enum pgt_entry entry, void *old_entry, void *new_entry) in move_pgt_entry() argument
608 moved = move_normal_pmd(pmc, old_entry, new_entry); in move_pgt_entry()
611 moved = move_normal_pud(pmc, old_entry, new_entry); in move_pgt_entry()
616 new_entry); in move_pgt_entry()
620 move_huge_pud(pmc, old_entry, new_entry); in move_pgt_entry()
/linux/fs/btrfs/
H A Dlru_cache.c107 struct btrfs_lru_cache_entry *new_entry, in btrfs_lru_cache_store() argument
110 const u64 key = new_entry->key; in btrfs_lru_cache_store()
121 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store()
126 if (match_entry(head, key, new_entry->gen) != NULL) in btrfs_lru_cache_store()
128 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store()
143 list_add_tail(&new_entry->lru_list, &cache->lru_list); in btrfs_lru_cache_store()
H A Dref-verify.c93 const struct block_entry *new_entry = rb_entry(new, struct block_entry, node); in block_entry_bytenr_cmp() local
95 return block_entry_bytenr_key_cmp(&new_entry->bytenr, existing); in block_entry_bytenr_cmp()
130 const struct root_entry *new_entry = rb_entry(new, struct root_entry, node); in root_entry_root_objectid_cmp() local
132 return root_entry_root_objectid_key_cmp(&new_entry->root_objectid, existing); in root_entry_root_objectid_cmp()
167 struct ref_entry *new_entry = rb_entry(new, struct ref_entry, node); in ref_entry_cmp() local
170 return comp_refs(new_entry, existing_entry); in ref_entry_cmp()
H A Dlru_cache.h65 struct btrfs_lru_cache_entry *new_entry,
/linux/security/tomoyo/
H A Ddomain.c31 int tomoyo_update_policy(struct tomoyo_acl_head *new_entry, const int size, in tomoyo_update_policy() argument
48 if (!check_duplicate(entry, new_entry)) in tomoyo_update_policy()
55 entry = tomoyo_commit_ok(new_entry, size); in tomoyo_update_policy()
92 int tomoyo_update_domain(struct tomoyo_acl_info *new_entry, const int size, in tomoyo_update_domain() argument
108 new_entry->cond = tomoyo_get_condition(param); in tomoyo_update_domain()
109 if (!new_entry->cond) in tomoyo_update_domain()
115 if (new_entry->cond->transit && in tomoyo_update_domain()
116 !(new_entry->type == TOMOYO_TYPE_PATH_ACL && in tomoyo_update_domain()
117 container_of(new_entry, struct tomoyo_path_acl, head) in tomoyo_update_domain()
127 if (!tomoyo_same_acl_head(entry, new_entry) || in tomoyo_update_domain()
[all …]
/linux/drivers/s390/cio/
H A Dqdio_debug.c63 struct qdio_dbf_entry *new_entry; in qdio_allocate_dbf() local
84 new_entry = kzalloc_obj(struct qdio_dbf_entry); in qdio_allocate_dbf()
85 if (!new_entry) { in qdio_allocate_dbf()
89 strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); in qdio_allocate_dbf()
90 new_entry->dbf_info = irq_ptr->debug_area; in qdio_allocate_dbf()
92 list_add(&new_entry->dbf_list, &qdio_dbf_list); in qdio_allocate_dbf()
/linux/tools/perf/util/
H A Drblist.c13 int rblist__add_node(struct rblist *rblist, const void *new_entry) in rblist__add_node() argument
24 rc = rblist->node_cmp(parent, new_entry); in rblist__add_node()
35 new_node = rblist->node_new(rblist, new_entry); in rblist__add_node()
H A Drblist.h27 struct rb_node *(*node_new)(struct rblist *rlist, const void *new_entry);
34 int rblist__add_node(struct rblist *rblist, const void *new_entry);
H A Dstrlist.c57 int strlist__add(struct strlist *slist, const char *new_entry) in strlist__add() argument
59 return rblist__add_node(&slist->rblist, new_entry); in strlist__add()
/linux/arch/x86/kvm/svm/
H A Davic.c386 u64 new_entry; in avic_init_backing_page()
401 BUILD_BUG_ON((AVIC_MAX_PHYSICAL_ID + 1) * sizeof(new_entry) > PAGE_SIZE || in avic_init_backing_page()
402 (X2AVIC_MAX_PHYSICAL_ID + 1) * sizeof(new_entry) > PAGE_SIZE); in avic_init_backing_page()
426 new_entry = avic_get_backing_page_address(svm) | in avic_init_backing_page()
428 svm->avic_physical_id_entry = new_entry; in avic_init_backing_page()
435 WRITE_ONCE(kvm_svm->avic_physical_id_table[id], new_entry); in avic_init_backing_page()
697 u32 *entry, new_entry; in avic_ldr_write()
704 new_entry = READ_ONCE(*entry); in avic_ldr_write()
705 new_entry &= ~AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK; in avic_ldr_write()
706 new_entry | in avic_ldr_write()
383 u64 new_entry; avic_init_backing_page() local
694 u32 *entry, new_entry; avic_ldr_write() local
[all...]
/linux/fs/smb/client/
H A Dreaddir.c463 char *new_entry; in nxt_dir_entry() local
470 new_entry = old_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 + in nxt_dir_entry()
479 new_entry = old_entry + next_offset; in nxt_dir_entry()
481 cifs_dbg(FYI, "new entry %p old entry %p\n", new_entry, old_entry); in nxt_dir_entry()
483 if (new_entry >= end_of_smb) { in nxt_dir_entry()
485 new_entry, end_of_smb, old_entry); in nxt_dir_entry()
488 (new_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 > end_of_smb)) in nxt_dir_entry()
490 (new_entry + sizeof(FILE_DIRECTORY_INFO) + 1 > end_of_smb))) { in nxt_dir_entry()
492 new_entry, end_of_smb); in nxt_dir_entry()
495 return new_entry; in nxt_dir_entry()
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dmcg.c155 struct mlx4_steer_index *new_entry; in new_steering_entry() local
165 new_entry = kzalloc_obj(*new_entry); in new_steering_entry()
166 if (!new_entry) in new_steering_entry()
169 INIT_LIST_HEAD(&new_entry->duplicates); in new_steering_entry()
170 new_entry->index = index; in new_steering_entry()
171 list_add_tail(&new_entry->list, &s_steer->steer_entries[steer]); in new_steering_entry()
184 list_add_tail(&dqp->list, &new_entry->duplicates); in new_steering_entry()
233 list_del(&new_entry->list); in new_steering_entry()
234 kfree(new_entry); in new_steering_entry()
1117 u8 new_entry = 0; in mlx4_qp_attach_common() local
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Daffinity.c587 bool new_entry = false; in hfi1_dev_affinity_init() local
608 new_entry = true; in hfi1_dev_affinity_init()
685 ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); in hfi1_dev_affinity_init()
689 if (new_entry) in hfi1_dev_affinity_init()
698 if (new_entry) in hfi1_dev_affinity_init()
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dcxgb4vf_main.c275 struct hash_mac_addr *new_entry, *entry; in cxgb4vf_change_mac() local
291 new_entry = kzalloc_obj(*new_entry); in cxgb4vf_change_mac()
292 if (!new_entry) in cxgb4vf_change_mac()
294 ether_addr_copy(new_entry->addr, addr); in cxgb4vf_change_mac()
295 new_entry->iface_mac = true; in cxgb4vf_change_mac()
296 list_add_tail(&new_entry->list, &adapter->mac_hlist); in cxgb4vf_change_mac()
945 struct hash_mac_addr *new_entry; in cxgb4vf_mac_sync() local
956 new_entry = kzalloc_obj(*new_entry, GFP_ATOMIC); in cxgb4vf_mac_sync()
957 if (!new_entry) in cxgb4vf_mac_sync()
959 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4vf_mac_sync()
[all …]
/linux/rust/kernel/
H A Dfirmware.rs320 pub const fn new_entry(self) -> Self { in new_entry() method
/linux/drivers/iommu/
H A Dvirtio-iommu.c972 struct iommu_resv_region *entry, *new_entry, *msi = NULL; in viommu_get_resv_regions() local
980 new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); in viommu_get_resv_regions()
981 if (!new_entry) in viommu_get_resv_regions()
983 list_add_tail(&new_entry->list, head); in viommu_get_resv_regions()
/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dspectrum_acl_tcam.c1193 struct mlxsw_sp_acl_tcam_entry *new_entry; in mlxsw_sp_acl_tcam_ventry_migrate() local
1202 new_entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, chunk); in mlxsw_sp_acl_tcam_ventry_migrate()
1203 if (IS_ERR(new_entry)) in mlxsw_sp_acl_tcam_ventry_migrate()
1204 return PTR_ERR(new_entry); in mlxsw_sp_acl_tcam_ventry_migrate()
1206 ventry->entry = new_entry; in mlxsw_sp_acl_tcam_ventry_migrate()
/linux/drivers/accel/habanalabs/common/
H A Ddebugfs.c1068 struct hl_debugfs_cfg_access_entry *new_entry; in check_if_cfg_access_and_log() local
1078 new_entry = &dbgfs_cfg_accesses->cfg_access_list[dbgfs_cfg_accesses->head]; in check_if_cfg_access_and_log()
1079 new_entry->seconds_since_epoch = ktime_get_real_seconds(); in check_if_cfg_access_and_log()
1080 new_entry->addr = addr; in check_if_cfg_access_and_log()
1081 new_entry->debugfs_type = access_type; in check_if_cfg_access_and_log()
1082 new_entry->valid = true; in check_if_cfg_access_and_log()
/linux/drivers/scsi/
H A Dhpsa.c1357 int entry, struct hpsa_scsi_dev_t *new_entry) in hpsa_scsi_update_entry() argument
1363 h->dev[entry]->raid_level = new_entry->raid_level; in hpsa_scsi_update_entry()
1368 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1371 if (new_entry->offload_config && new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry()
1380 h->dev[entry]->raid_map = new_entry->raid_map; in hpsa_scsi_update_entry()
1381 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1383 if (new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry()
1384 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1387 h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled; in hpsa_scsi_update_entry()
1388 h->dev[entry]->offload_config = new_entry->offload_config; in hpsa_scsi_update_entry()
[all …]
/linux/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_npc.c2949 u16 old_entry, new_entry; in rvu_mbox_handler_npc_mcam_shift_entry() local
2963 new_entry = req->new_entry[index]; in rvu_mbox_handler_npc_mcam_shift_entry()
2972 rc = npc_mcam_verify_entry(mcam, pcifunc, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry()
2977 if (mcam->entry2cntr_map[new_entry] != NPC_MCAM_INVALID_MAP) { in rvu_mbox_handler_npc_mcam_shift_entry()
2983 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, false); in rvu_mbox_handler_npc_mcam_shift_entry()
2986 npc_copy_mcam_entry(rvu, mcam, blkaddr, old_entry, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry()
2994 new_entry, cntr); in rvu_mbox_handler_npc_mcam_shift_entry()
2998 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, true); in rvu_mbox_handler_npc_mcam_shift_entry()
/linux/drivers/md/dm-vdo/
H A Drecovery-journal.c1341 struct recovery_journal_entry new_entry; in add_queued_recovery_entries() local
1349 new_entry = (struct recovery_journal_entry) { in add_queued_recovery_entries()
1361 *packed_entry = vdo_pack_recovery_journal_entry(&new_entry); in add_queued_recovery_entries()

12