| /linux/fs/ext4/ |
| H A D | block_validity.c | 73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local 88 new_entry = kmem_cache_alloc(ext4_system_zone_cachep, in add_system_zone() 90 if (!new_entry) in add_system_zone() 92 new_entry->start_blk = start_blk; in add_system_zone() 93 new_entry->count = count; in add_system_zone() 94 new_entry->ino = ino; in add_system_zone() 95 new_node = &new_entry->node; in add_system_zone() 104 if (can_merge(entry, new_entry)) { in add_system_zone() 105 new_entry->start_blk = entry->start_blk; in add_system_zone() 106 new_entry->count += entry->count; in add_system_zone() [all …]
|
| H A D | mballoc.c | 6440 struct ext4_free_data *new_entry) in ext4_mb_free_metadata() argument 6444 ext4_grpblk_t clusters = new_entry->efd_count; in ext4_mb_free_metadata() 6457 new_node = &new_entry->efd_node; in ext4_mb_free_metadata() 6458 cluster = new_entry->efd_start_cluster; in ext4_mb_free_metadata() 6481 kmem_cache_free(ext4_free_data_cachep, new_entry); in ext4_mb_free_metadata() 6491 if (ext4_freed_extents_can_be_merged(new_entry, entry)) { in ext4_mb_free_metadata() 6493 entry->efd_count += new_entry->efd_count; in ext4_mb_free_metadata() 6494 kmem_cache_free(ext4_free_data_cachep, new_entry); in ext4_mb_free_metadata() 6498 if (ext4_freed_extents_can_be_merged(entry, new_entry)) { in ext4_mb_free_metadata() 6499 entry->efd_count += new_entry->efd_count; in ext4_mb_free_metadata() [all …]
|
| /linux/drivers/iommu/generic_pt/ |
| H A D | pt_fmt_defaults.h | 224 u64 new_entry; in pt_set_sw_bit_release() local 227 new_entry = old_entry | pt_sw_bit(bitnr); in pt_set_sw_bit_release() 228 } while (!try_cmpxchg64_release(entryp, &old_entry, new_entry)); in pt_set_sw_bit_release() 229 pts->entry = new_entry; in pt_set_sw_bit_release() 236 u32 new_entry; in pt_set_sw_bit_release() local 239 new_entry = old_entry | pt_sw_bit(bitnr); in pt_set_sw_bit_release() 240 } while (!try_cmpxchg_release(entryp, &old_entry, new_entry)); in pt_set_sw_bit_release() 241 pts->entry = new_entry; in pt_set_sw_bit_release()
|
| /linux/mm/ |
| H A D | shmem_quota.c | 170 struct quota_id *new_entry, *entry; in shmem_acquire_dquot() local 193 new_entry = kzalloc_obj(struct quota_id, GFP_NOFS); in shmem_acquire_dquot() 194 if (!new_entry) { in shmem_acquire_dquot() 199 new_entry->id = id; in shmem_acquire_dquot() 201 new_entry->bhardlimit = sbinfo->qlimits.usrquota_bhardlimit; in shmem_acquire_dquot() 202 new_entry->ihardlimit = sbinfo->qlimits.usrquota_ihardlimit; in shmem_acquire_dquot() 204 new_entry->bhardlimit = sbinfo->qlimits.grpquota_bhardlimit; in shmem_acquire_dquot() 205 new_entry->ihardlimit = sbinfo->qlimits.grpquota_ihardlimit; in shmem_acquire_dquot() 208 new_node = &new_entry->node; in shmem_acquire_dquot() 211 entry = new_entry; in shmem_acquire_dquot()
|
| H A D | mremap.c | 597 enum pgt_entry entry, void *old_entry, void *new_entry) in move_pgt_entry() argument 608 moved = move_normal_pmd(pmc, old_entry, new_entry); in move_pgt_entry() 611 moved = move_normal_pud(pmc, old_entry, new_entry); in move_pgt_entry() 616 new_entry); in move_pgt_entry() 620 move_huge_pud(pmc, old_entry, new_entry); in move_pgt_entry()
|
| /linux/fs/btrfs/ |
| H A D | lru_cache.c | 107 struct btrfs_lru_cache_entry *new_entry, in btrfs_lru_cache_store() argument 110 const u64 key = new_entry->key; in btrfs_lru_cache_store() 121 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store() 126 if (match_entry(head, key, new_entry->gen) != NULL) in btrfs_lru_cache_store() 128 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store() 143 list_add_tail(&new_entry->lru_list, &cache->lru_list); in btrfs_lru_cache_store()
|
| H A D | ref-verify.c | 93 const struct block_entry *new_entry = rb_entry(new, struct block_entry, node); in block_entry_bytenr_cmp() local 95 return block_entry_bytenr_key_cmp(&new_entry->bytenr, existing); in block_entry_bytenr_cmp() 130 const struct root_entry *new_entry = rb_entry(new, struct root_entry, node); in root_entry_root_objectid_cmp() local 132 return root_entry_root_objectid_key_cmp(&new_entry->root_objectid, existing); in root_entry_root_objectid_cmp() 167 struct ref_entry *new_entry = rb_entry(new, struct ref_entry, node); in ref_entry_cmp() local 170 return comp_refs(new_entry, existing_entry); in ref_entry_cmp()
|
| H A D | lru_cache.h | 65 struct btrfs_lru_cache_entry *new_entry,
|
| /linux/security/tomoyo/ |
| H A D | domain.c | 31 int tomoyo_update_policy(struct tomoyo_acl_head *new_entry, const int size, in tomoyo_update_policy() argument 48 if (!check_duplicate(entry, new_entry)) in tomoyo_update_policy() 55 entry = tomoyo_commit_ok(new_entry, size); in tomoyo_update_policy() 92 int tomoyo_update_domain(struct tomoyo_acl_info *new_entry, const int size, in tomoyo_update_domain() argument 108 new_entry->cond = tomoyo_get_condition(param); in tomoyo_update_domain() 109 if (!new_entry->cond) in tomoyo_update_domain() 115 if (new_entry->cond->transit && in tomoyo_update_domain() 116 !(new_entry->type == TOMOYO_TYPE_PATH_ACL && in tomoyo_update_domain() 117 container_of(new_entry, struct tomoyo_path_acl, head) in tomoyo_update_domain() 127 if (!tomoyo_same_acl_head(entry, new_entry) || in tomoyo_update_domain() [all …]
|
| /linux/drivers/s390/cio/ |
| H A D | qdio_debug.c | 63 struct qdio_dbf_entry *new_entry; in qdio_allocate_dbf() local 84 new_entry = kzalloc_obj(struct qdio_dbf_entry); in qdio_allocate_dbf() 85 if (!new_entry) { in qdio_allocate_dbf() 89 strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); in qdio_allocate_dbf() 90 new_entry->dbf_info = irq_ptr->debug_area; in qdio_allocate_dbf() 92 list_add(&new_entry->dbf_list, &qdio_dbf_list); in qdio_allocate_dbf()
|
| /linux/tools/perf/util/ |
| H A D | rblist.c | 13 int rblist__add_node(struct rblist *rblist, const void *new_entry) in rblist__add_node() argument 24 rc = rblist->node_cmp(parent, new_entry); in rblist__add_node() 35 new_node = rblist->node_new(rblist, new_entry); in rblist__add_node()
|
| H A D | rblist.h | 27 struct rb_node *(*node_new)(struct rblist *rlist, const void *new_entry); 34 int rblist__add_node(struct rblist *rblist, const void *new_entry);
|
| H A D | strlist.c | 57 int strlist__add(struct strlist *slist, const char *new_entry) in strlist__add() argument 59 return rblist__add_node(&slist->rblist, new_entry); in strlist__add()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | avic.c | 386 u64 new_entry; in avic_init_backing_page() 401 BUILD_BUG_ON((AVIC_MAX_PHYSICAL_ID + 1) * sizeof(new_entry) > PAGE_SIZE || in avic_init_backing_page() 402 (X2AVIC_MAX_PHYSICAL_ID + 1) * sizeof(new_entry) > PAGE_SIZE); in avic_init_backing_page() 426 new_entry = avic_get_backing_page_address(svm) | in avic_init_backing_page() 428 svm->avic_physical_id_entry = new_entry; in avic_init_backing_page() 435 WRITE_ONCE(kvm_svm->avic_physical_id_table[id], new_entry); in avic_init_backing_page() 697 u32 *entry, new_entry; in avic_ldr_write() 704 new_entry = READ_ONCE(*entry); in avic_ldr_write() 705 new_entry &= ~AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK; in avic_ldr_write() 706 new_entry | in avic_ldr_write() 383 u64 new_entry; avic_init_backing_page() local 694 u32 *entry, new_entry; avic_ldr_write() local [all...] |
| /linux/fs/smb/client/ |
| H A D | readdir.c | 463 char *new_entry; in nxt_dir_entry() local 470 new_entry = old_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 + in nxt_dir_entry() 479 new_entry = old_entry + next_offset; in nxt_dir_entry() 481 cifs_dbg(FYI, "new entry %p old entry %p\n", new_entry, old_entry); in nxt_dir_entry() 483 if (new_entry >= end_of_smb) { in nxt_dir_entry() 485 new_entry, end_of_smb, old_entry); in nxt_dir_entry() 488 (new_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 > end_of_smb)) in nxt_dir_entry() 490 (new_entry + sizeof(FILE_DIRECTORY_INFO) + 1 > end_of_smb))) { in nxt_dir_entry() 492 new_entry, end_of_smb); in nxt_dir_entry() 495 return new_entry; in nxt_dir_entry()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | mcg.c | 155 struct mlx4_steer_index *new_entry; in new_steering_entry() local 165 new_entry = kzalloc_obj(*new_entry); in new_steering_entry() 166 if (!new_entry) in new_steering_entry() 169 INIT_LIST_HEAD(&new_entry->duplicates); in new_steering_entry() 170 new_entry->index = index; in new_steering_entry() 171 list_add_tail(&new_entry->list, &s_steer->steer_entries[steer]); in new_steering_entry() 184 list_add_tail(&dqp->list, &new_entry->duplicates); in new_steering_entry() 233 list_del(&new_entry->list); in new_steering_entry() 234 kfree(new_entry); in new_steering_entry() 1117 u8 new_entry = 0; in mlx4_qp_attach_common() local [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | affinity.c | 587 bool new_entry = false; in hfi1_dev_affinity_init() local 608 new_entry = true; in hfi1_dev_affinity_init() 685 ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); in hfi1_dev_affinity_init() 689 if (new_entry) in hfi1_dev_affinity_init() 698 if (new_entry) in hfi1_dev_affinity_init()
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | cxgb4vf_main.c | 275 struct hash_mac_addr *new_entry, *entry; in cxgb4vf_change_mac() local 291 new_entry = kzalloc_obj(*new_entry); in cxgb4vf_change_mac() 292 if (!new_entry) in cxgb4vf_change_mac() 294 ether_addr_copy(new_entry->addr, addr); in cxgb4vf_change_mac() 295 new_entry->iface_mac = true; in cxgb4vf_change_mac() 296 list_add_tail(&new_entry->list, &adapter->mac_hlist); in cxgb4vf_change_mac() 945 struct hash_mac_addr *new_entry; in cxgb4vf_mac_sync() local 956 new_entry = kzalloc_obj(*new_entry, GFP_ATOMIC); in cxgb4vf_mac_sync() 957 if (!new_entry) in cxgb4vf_mac_sync() 959 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4vf_mac_sync() [all …]
|
| /linux/rust/kernel/ |
| H A D | firmware.rs | 320 pub const fn new_entry(self) -> Self { in new_entry() method
|
| /linux/drivers/iommu/ |
| H A D | virtio-iommu.c | 972 struct iommu_resv_region *entry, *new_entry, *msi = NULL; in viommu_get_resv_regions() local 980 new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); in viommu_get_resv_regions() 981 if (!new_entry) in viommu_get_resv_regions() 983 list_add_tail(&new_entry->list, head); in viommu_get_resv_regions()
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum_acl_tcam.c | 1193 struct mlxsw_sp_acl_tcam_entry *new_entry; in mlxsw_sp_acl_tcam_ventry_migrate() local 1202 new_entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, chunk); in mlxsw_sp_acl_tcam_ventry_migrate() 1203 if (IS_ERR(new_entry)) in mlxsw_sp_acl_tcam_ventry_migrate() 1204 return PTR_ERR(new_entry); in mlxsw_sp_acl_tcam_ventry_migrate() 1206 ventry->entry = new_entry; in mlxsw_sp_acl_tcam_ventry_migrate()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | debugfs.c | 1068 struct hl_debugfs_cfg_access_entry *new_entry; in check_if_cfg_access_and_log() local 1078 new_entry = &dbgfs_cfg_accesses->cfg_access_list[dbgfs_cfg_accesses->head]; in check_if_cfg_access_and_log() 1079 new_entry->seconds_since_epoch = ktime_get_real_seconds(); in check_if_cfg_access_and_log() 1080 new_entry->addr = addr; in check_if_cfg_access_and_log() 1081 new_entry->debugfs_type = access_type; in check_if_cfg_access_and_log() 1082 new_entry->valid = true; in check_if_cfg_access_and_log()
|
| /linux/drivers/scsi/ |
| H A D | hpsa.c | 1357 int entry, struct hpsa_scsi_dev_t *new_entry) in hpsa_scsi_update_entry() argument 1363 h->dev[entry]->raid_level = new_entry->raid_level; in hpsa_scsi_update_entry() 1368 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1371 if (new_entry->offload_config && new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1380 h->dev[entry]->raid_map = new_entry->raid_map; in hpsa_scsi_update_entry() 1381 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1383 if (new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1384 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1387 h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled; in hpsa_scsi_update_entry() 1388 h->dev[entry]->offload_config = new_entry->offload_config; in hpsa_scsi_update_entry() [all …]
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu_npc.c | 2949 u16 old_entry, new_entry; in rvu_mbox_handler_npc_mcam_shift_entry() local 2963 new_entry = req->new_entry[index]; in rvu_mbox_handler_npc_mcam_shift_entry() 2972 rc = npc_mcam_verify_entry(mcam, pcifunc, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry() 2977 if (mcam->entry2cntr_map[new_entry] != NPC_MCAM_INVALID_MAP) { in rvu_mbox_handler_npc_mcam_shift_entry() 2983 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, false); in rvu_mbox_handler_npc_mcam_shift_entry() 2986 npc_copy_mcam_entry(rvu, mcam, blkaddr, old_entry, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry() 2994 new_entry, cntr); in rvu_mbox_handler_npc_mcam_shift_entry() 2998 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, true); in rvu_mbox_handler_npc_mcam_shift_entry()
|
| /linux/drivers/md/dm-vdo/ |
| H A D | recovery-journal.c | 1341 struct recovery_journal_entry new_entry; in add_queued_recovery_entries() local 1349 new_entry = (struct recovery_journal_entry) { in add_queued_recovery_entries() 1361 *packed_entry = vdo_pack_recovery_journal_entry(&new_entry); in add_queued_recovery_entries()
|