/linux/drivers/md/bcache/ |
H A D | extents.c | 229 struct bkey *insert, in bch_btree_ptr_insert_fixup() argument 235 if (!KEY_OFFSET(insert)) in bch_btree_ptr_insert_fixup() 331 struct bkey *insert, in bch_extent_insert_fixup() argument 340 BUG_ON(!KEY_OFFSET(insert)); in bch_extent_insert_fixup() 341 BUG_ON(!KEY_SIZE(insert)); in bch_extent_insert_fixup() 349 if (bkey_cmp(&START_KEY(k), insert) >= 0) { in bch_extent_insert_fixup() 356 if (bkey_cmp(k, &START_KEY(insert)) <= 0) in bch_extent_insert_fixup() 385 if (KEY_START(k) > KEY_START(insert) + sectors_found) in bch_extent_insert_fixup() 400 sectors_found = KEY_OFFSET(k) - KEY_START(insert); in bch_extent_insert_fixup() 403 if (bkey_cmp(insert, k) < 0 && in bch_extent_insert_fixup() [all …]
|
/linux/fs/bcachefs/ |
H A D | data_update.c | 96 struct bkey_i *insert) in trace_move_extent_finish2() argument 111 bch2_bkey_val_to_text(&buf, c, bkey_i_to_s_c(insert)); in trace_move_extent_finish2() 121 struct bkey_i *insert, in trace_move_extent_fail2() argument 134 if (insert) { in trace_move_extent_fail2() 142 (ptr = bch2_extent_has_ptr(old, p, bkey_i_to_s(insert))) && in trace_move_extent_fail2() 164 if (insert) { in trace_move_extent_fail2() 166 bch2_bkey_val_to_text(&buf, c, bkey_i_to_s_c(insert)); in trace_move_extent_fail2() 195 struct bkey_i *insert = NULL; in __bch2_data_update_index_update() local 223 insert = _insert.k; in __bch2_data_update_index_update() 229 bch2_cut_front(iter.pos, insert); in __bch2_data_update_index_update() [all …]
|
H A D | extent_update.c | 107 struct bkey_i *insert, in bch2_extent_atomic_end() argument 119 *end = insert->k.p; in bch2_extent_atomic_end() 124 ret = count_iters_for_insert(trans, bkey_i_to_s_c(insert), 0, end, in bch2_extent_atomic_end() 131 for_each_btree_key_max_continue_norestart(copy, insert->k.p, 0, k, ret) { in bch2_extent_atomic_end() 134 if (bkey_gt(bkey_start_pos(&insert->k), bkey_start_pos(k.k))) in bch2_extent_atomic_end() 135 offset = bkey_start_offset(&insert->k) - in bch2_extent_atomic_end() 139 switch (bch2_extent_overlap(&insert->k, k.k)) { in bch2_extent_atomic_end()
|
H A D | str_hash.h | 260 struct bkey_i *insert, in bch2_hash_set_or_get_in_snapshot() argument 269 SPOS(insert->k.p.inode, in bch2_hash_set_or_get_in_snapshot() 270 desc.hash_bkey(info, bkey_i_to_s_c(insert)), in bch2_hash_set_or_get_in_snapshot() 272 POS(insert->k.p.inode, U64_MAX), in bch2_hash_set_or_get_in_snapshot() 275 if (!desc.cmp_bkey(k, bkey_i_to_s_c(insert))) in bch2_hash_set_or_get_in_snapshot() 307 insert->k.p = iter->pos; in bch2_hash_set_or_get_in_snapshot() 308 ret = bch2_trans_update(trans, iter, insert, flags); in bch2_hash_set_or_get_in_snapshot() 319 struct bkey_i *insert, in bch2_hash_set_in_snapshot() argument 324 snapshot, insert, flags); in bch2_hash_set_in_snapshot() 341 struct bkey_i *insert, in bch2_hash_set() argument [all …]
|
H A D | btree_update.c | 34 struct bkey_i **insert, in extent_front_merge() argument 49 if (!bch2_bkey_merge(c, bkey_i_to_s(update), bkey_i_to_s_c(*insert))) in extent_front_merge() 53 bch2_key_has_snapshot_overwrites(trans, iter->btree_id, (*insert)->k.p); in extent_front_merge() 63 *insert = update; in extent_front_merge() 69 struct bkey_i *insert, in extent_back_merge() argument 78 ret = bch2_key_has_snapshot_overwrites(trans, iter->btree_id, insert->k.p) ?: in extent_back_merge() 85 bch2_bkey_merge(c, bkey_i_to_s(insert), k); in extent_back_merge() 287 struct bkey_i *insert, in bch2_trans_update_extent() argument 295 bch2_trans_iter_init(trans, &iter, btree_id, bkey_start_pos(&insert->k), in bch2_trans_update_extent() 299 k = bch2_btree_iter_peek_max(&iter, POS(insert->k.p.inode, U64_MAX)); in bch2_trans_update_extent() [all …]
|
H A D | io_misc.c | 369 bool insert = shift > 0; in __bch2_resume_logged_op_finsert() local 394 if (insert) { in __bch2_resume_logged_op_finsert() 430 k = insert in __bch2_resume_logged_op_finsert() 445 if (insert && in __bch2_resume_logged_op_finsert() 464 op->v.pos = cpu_to_le64(insert ? bkey_start_offset(&delete.k) : delete.k.p.offset); in __bch2_resume_logged_op_finsert() 484 if (!insert) { in __bch2_resume_logged_op_finsert() 512 u64 offset, u64 len, bool insert, in bch2_fcollapse_finsert() argument 516 s64 shift = insert ? len : -len; in bch2_fcollapse_finsert() 523 op.v.pos = cpu_to_le64(insert ? U64_MAX : offset); in bch2_fcollapse_finsert()
|
H A D | btree_trans_commit.c | 155 struct bkey_i *insert) in bch2_btree_bset_insert_key() argument 162 EBUG_ON(bkey_deleted(&insert->k) && bkey_val_u64s(&insert->k)); in bch2_btree_bset_insert_key() 163 EBUG_ON(bpos_lt(insert->k.p, b->data->min_key)); in bch2_btree_bset_insert_key() 164 EBUG_ON(bpos_gt(insert->k.p, b->data->max_key)); in bch2_btree_bset_insert_key() 165 EBUG_ON(insert->k.u64s > bch2_btree_keys_u64s_remaining(b)); in bch2_btree_bset_insert_key() 166 EBUG_ON(!b->c.level && !bpos_eq(insert->k.p, path->pos)); in bch2_btree_bset_insert_key() 169 if (k && bkey_cmp_left_packed(b, k, &insert->k.p)) in bch2_btree_bset_insert_key() 176 if (bkey_deleted(&insert->k) && !k) in bch2_btree_bset_insert_key() 179 if (bkey_deleted(&insert->k)) { in bch2_btree_bset_insert_key() 185 push_whiteout(b, insert->k.p); in bch2_btree_bset_insert_key() [all …]
|
H A D | backpointers.h | 103 bool insert) in bch2_bucket_backpointer_mod() argument 106 return bch2_bucket_backpointer_mod_nowritebuffer(trans, orig_k, bp, insert); in bch2_bucket_backpointer_mod() 108 if (!insert) { in bch2_bucket_backpointer_mod()
|
H A D | bset.c | 241 struct bkey_packed *insert, unsigned clobber_u64s) in bch2_verify_insert_pos() argument 250 bkey_iter_cmp(b, prev, insert) > 0); in bch2_verify_insert_pos() 253 bkey_iter_cmp(b, prev, insert) > 0) { in bch2_verify_insert_pos() 255 struct bkey k2 = bkey_unpack_key(b, insert); in bch2_verify_insert_pos() 269 bkey_iter_cmp(b, insert, next) > 0); in bch2_verify_insert_pos() 272 bkey_iter_cmp(b, insert, next) > 0) { in bch2_verify_insert_pos() 273 struct bkey k1 = bkey_unpack_key(b, insert); in bch2_verify_insert_pos() 981 struct bkey_i *insert, in bch2_bset_insert() argument 986 struct bkey_packed packed, *src = bkey_to_packed(insert); in bch2_bset_insert() 989 bch2_verify_insert_pos(b, where, bkey_to_packed(insert), clobber_u64s); in bch2_bset_insert() [all …]
|
H A D | btree_update_interior.c | 910 struct bkey_i *insert = &b->key; in btree_update_updated_root() local 915 BUG_ON(as->journal_u64s + jset_u64s(insert->k.u64s) > in btree_update_updated_root() 922 insert, insert->k.u64s); in btree_update_updated_root() 1341 struct bkey_i *insert) in bch2_insert_fixup_btree_ptr() argument 1348 BUG_ON(insert->k.type == KEY_TYPE_btree_ptr_v2 && in bch2_insert_fixup_btree_ptr() 1349 !btree_ptr_sectors_written(bkey_i_to_s_c(insert))); in bch2_insert_fixup_btree_ptr() 1352 bch2_journal_key_overwritten(c, b->c.btree_id, b->c.level, insert->k.p); in bch2_insert_fixup_btree_ptr() 1360 if (bch2_bkey_validate(c, bkey_i_to_s_c(insert), from) ?: in bch2_insert_fixup_btree_ptr() 1361 bch2_bkey_in_btree_node(c, b, bkey_i_to_s_c(insert), from)) { in bch2_insert_fixup_btree_ptr() 1366 BUG_ON(as->journal_u64s + jset_u64s(insert->k.u64s) > in bch2_insert_fixup_btree_ptr() [all …]
|
H A D | buckets.c | 545 bool insert) in __mark_pointer() argument 555 if (insert) in __mark_pointer() 568 bool insert = !(flags & BTREE_TRIGGER_overwrite); in bch2_trigger_pointer() local 575 *sectors = insert ? bp.v.bucket_len : -(s64) bp.v.bucket_len; in bch2_trigger_pointer() 579 if (insert && p.ptr.dev != BCH_SB_MEMBER_INVALID) in bch2_trigger_pointer() 589 __mark_pointer(trans, ca, k, &p, *sectors, bp.v.data_type, &a->v, insert); in bch2_trigger_pointer() 594 ret = bch2_bucket_backpointer_mod(trans, k, &bp, insert); in bch2_trigger_pointer() 611 ret = __mark_pointer(trans, ca, k, &p, *sectors, bp.v.data_type, &new, insert); in bch2_trigger_pointer() 822 bool insert = !(flags & BTREE_TRIGGER_overwrite); in __trigger_extent() local 828 insert ? 1 : -1, in __trigger_extent() [all …]
|
/linux/virt/kvm/ |
H A D | coalesced_mmio.c | 49 __u32 insert; in coalesced_mmio_write() local 62 insert = READ_ONCE(ring->last); in coalesced_mmio_write() 63 if (insert >= KVM_COALESCED_MMIO_MAX || in coalesced_mmio_write() 64 (insert + 1) % KVM_COALESCED_MMIO_MAX == READ_ONCE(ring->first)) { in coalesced_mmio_write() 71 ring->coalesced_mmio[insert].phys_addr = addr; in coalesced_mmio_write() 72 ring->coalesced_mmio[insert].len = len; in coalesced_mmio_write() 73 memcpy(ring->coalesced_mmio[insert].data, val, len); in coalesced_mmio_write() 74 ring->coalesced_mmio[insert].pio = dev->zone.pio; in coalesced_mmio_write() 76 ring->last = (insert + 1) % KVM_COALESCED_MMIO_MAX; in coalesced_mmio_write()
|
/linux/tools/testing/selftests/tc-testing/plugin-lib/ |
H A D | valgrindPlugin.py | 84 cmdlist.insert(0, '--track-origins=yes') 85 cmdlist.insert(0, '--show-leak-kinds=definite,indirect') 86 cmdlist.insert(0, '--leak-check=full') 87 cmdlist.insert(0, '--log-file=vgnd-{}.log'.format(self.args.testid)) 88 cmdlist.insert(0, '-v') # ask for summary of non-leak errors 89 cmdlist.insert(0, ENVIR['VALGRIND_BIN'])
|
H A D | nsPlugin.py | 102 cmdlist.insert(0, self.args.NAMES['NS']) 103 cmdlist.insert(0, 'exec') 104 cmdlist.insert(0, 'netns') 105 cmdlist.insert(0, self.args.NAMES['IP'])
|
/linux/drivers/accessibility/speakup/ |
H A D | DefaultKeyAssignments | 6 We have remapped the insert or zero key on the keypad to act as a 8 InsKeyPad-period means hold down the insert key like a shift key and 31 InsKeyPad-9 Move reading cursor to top of screen (insert pgup) 32 InsKeyPad-3 Move reading cursor to bottom of screen (insert pgdn) 33 InsKeyPad-7 Move reading cursor to left edge of screen (insert home) 34 InsKeyPad-1 Move reading cursor to right edge of screen (insert end)
|
/linux/tools/perf/util/ |
H A D | comm.c | 165 int insert = comm_strs->num_strs; /* Default to inserting at the end. */ in comm_strs__findnew() local 175 insert = mid; in comm_strs__findnew() 178 memmove(&comm_strs->strs[insert + 1], &comm_strs->strs[insert], in comm_strs__findnew() 179 (comm_strs->num_strs - insert) * sizeof(struct comm_str *)); in comm_strs__findnew() 181 comm_strs->strs[insert] = result; in comm_strs__findnew()
|
H A D | dsos.c | 213 int insert = dsos->cnt; /* Default to inserting at the end. */ in __dsos__add() local 223 insert = mid; in __dsos__add() 226 memmove(&dsos->dsos[insert + 1], &dsos->dsos[insert], in __dsos__add() 227 (dsos->cnt - insert) * sizeof(struct dso *)); in __dsos__add() 229 dsos->dsos[insert] = dso__get(dso); in __dsos__add()
|
/linux/fs/notify/ |
H A D | notification.c | 85 void (*insert)(struct fsnotify_group *, in fsnotify_insert_event() 123 if (insert) in fsnotify_insert_event() 124 insert(group, event); in fsnotify_insert_event()
|
/linux/net/rds/ |
H A D | cong.c | 105 struct rds_cong_map *insert) in rds_cong_tree_walk() argument 126 if (insert) { in rds_cong_tree_walk() 127 rb_link_node(&insert->m_rb_node, parent, p); in rds_cong_tree_walk() 128 rb_insert_color(&insert->m_rb_node, &rds_cong_tree); in rds_cong_tree_walk()
|
/linux/Documentation/sphinx/ |
H A D | rstFlatTable.py | 130 tableNode.insert(0, title) 254 self.rows[y].insert(x+c+1, None) 262 self.rows[y+r+1].insert(x+c, None) 345 cellElements.insert(0, target)
|
/linux/drivers/net/can/dev/ |
H A D | rx-offload.c | 80 struct sk_buff *pos, *insert = NULL; in __skb_queue_add_sort() local 97 insert = pos; in __skb_queue_add_sort() 100 if (!insert) in __skb_queue_add_sort() 103 __skb_queue_after(head, insert, new); in __skb_queue_add_sort()
|
/linux/rust/macros/ |
H A D | vtable.rs | 48 consts.insert(const_name); in vtable() 77 consts.insert(gen_const_name); in vtable()
|
/linux/Documentation/userspace-api/media/v4l/ |
H A D | selection-api-intro.rst | 9 can insert the image into larger one. Some video output devices can crop 10 part of an input image, scale it up or down and insert it at an
|
/linux/Documentation/ |
H A D | conf.py | 44 sys.path.insert(0, os.path.abspath('sphinx')) 361 html_sidebars['**'].insert(0, 'about.html')
|
/linux/rust/kernel/ |
H A D | rbtree.rs | 286 Ok(self.insert(RBTreeNode::new(key, value, flags)?)) in try_create_and_insert() 295 pub fn insert(&mut self, node: RBTreeNode<K, V>) -> Option<RBTreeNode<K, V>> { in insert() function 299 entry.insert(node); in insert() 1168 fn insert(self, node: RBTreeNode<K, V>) -> &'a mut V { in insert() method 1190 pub fn insert(self, value: V, reservation: RBTreeNodeReservation<K, V>) -> &'a mut V { in insert() method 1191 self.raw.insert(reservation.into_node(self.key, value)) in insert()
|