Home
last modified time | relevance | path

Searched full:k (Results 1 – 25 of 2765) sorted by relevance

12345678910>>...111

/linux/include/uapi/linux/
H A Dkeyboard.h45 #define K(t,v) (((t)<<8)|(v)) macro
49 #define K_F1 K(KT_FN,0)
50 #define K_F2 K(KT_FN,1)
51 #define K_F3 K(KT_FN,2)
52 #define K_F4 K(KT_FN,3)
53 #define K_F5 K(KT_FN,4)
54 #define K_F6 K(KT_FN,5)
55 #define K_F7 K(KT_FN,6)
56 #define K_F8 K(KT_FN,7)
57 #define K_F9 K(KT_FN,8)
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/calcs/
H A Ddcn_calc_auto.c42 int k; in scaler_settings_calculation() local
43 for (k = 0; k <= v->number_of_active_planes - 1; k++) { in scaler_settings_calculation()
45 if (v->source_scan[k] == dcn_bw_hor) { in scaler_settings_calculation()
46 v->h_ratio[k] = v->viewport_width[k] / v->scaler_rec_out_width[k]; in scaler_settings_calculation()
47 v->v_ratio[k] = v->viewport_height[k] / v->scaler_recout_height[k]; in scaler_settings_calculation()
50 v->h_ratio[k] = v->viewport_height[k] / v->scaler_rec_out_width[k]; in scaler_settings_calculation()
51 v->v_ratio[k] = v->viewport_width[k] / v->scaler_recout_height[k]; in scaler_settings_calculation()
55 if (v->source_scan[k] == dcn_bw_hor) { in scaler_settings_calculation()
56 …v->h_ratio[k] =dcn_bw_max2(v->viewport_width[k] / v->scaler_rec_out_width[k], v->viewport_height[k in scaler_settings_calculation()
59 …v->h_ratio[k] =dcn_bw_max2(v->viewport_height[k] / v->scaler_rec_out_width[k], v->viewport_width[k in scaler_settings_calculation()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn30/
H A Ddisplay_mode_vba_30.c82 unsigned int k,
745 unsigned int k, in CalculatePrefetchSchedule() argument
830 v->DynamicMetadataTransmittedBytes[k], in CalculatePrefetchSchedule()
831 v->DynamicMetadataLinesBeforeActiveRequired[k], in CalculatePrefetchSchedule()
844 v->Tdmdl[k] = TWait + Tvm_trips + trip_to_mem; in CalculatePrefetchSchedule()
846 v->Tdmdl[k] = TWait + UrgentExtraLatency; in CalculatePrefetchSchedule()
849 if (v->DynamicMetadataEnable[k] == true) { in CalculatePrefetchSchedule()
850 if (VStartup * LineTime < Tsetup + v->Tdmdl[k] + Tdmbf + Tdmec + Tdmsks) { in CalculatePrefetchSchedule()
858 dml_print("DML: Tdmdl: %fus - time for fabric to become ready and fetch dmd \n", v->Tdmdl[k]); in CalculatePrefetchSchedule()
864 …v->Tdmdl_vm[k] = (v->DynamicMetadataEnable[k] == true && v->DynamicMetadataVMEnabled == true && v-… in CalculatePrefetchSchedule()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/
H A Ddisplay_mode_vba_32.c62 unsigned int j, k; in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation() local
85 for (k = 0; k < mode_lib->vba.NumberOfActiveSurfaces; ++k) { in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
86 if (mode_lib->vba.WritebackEnable[k]) { in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
89 mode_lib->vba.WritebackPixelFormat[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
90 mode_lib->vba.PixelClock[k], mode_lib->vba.WritebackHRatio[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
91 mode_lib->vba.WritebackVRatio[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
92 mode_lib->vba.WritebackHTaps[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
93 mode_lib->vba.WritebackVTaps[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
94 mode_lib->vba.WritebackSourceWidth[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
95 mode_lib->vba.WritebackDestinationWidth[k], in DISPCLKDPPCLKDCFCLKDeepSleepPrefetchParametersWatermarksAndPerformanceCalculation()
[all …]
H A Ddisplay_mode_vba_util_32.c468 unsigned int k; in dml32_CalculateSwathAndDETConfiguration() local
515 for (k = 0; k < NumberOfActiveSurfaces; ++k) { in dml32_CalculateSwathAndDETConfiguration()
516 …RoundedUpMaxSwathSizeBytesY[k] = swath_width_luma_ub[k] * BytePerPixDETY[k] * MaximumSwathHeightY[ in dml32_CalculateSwathAndDETConfiguration()
517 …RoundedUpMaxSwathSizeBytesC[k] = swath_width_chroma_ub[k] * BytePerPixDETC[k] * MaximumSwathHeight… in dml32_CalculateSwathAndDETConfiguration()
519 dml_print("DML::%s: k=%0d DPPPerSurface = %d\n", __func__, k, DPPPerSurface[k]); in dml32_CalculateSwathAndDETConfiguration()
520 dml_print("DML::%s: k=%0d swath_width_luma_ub = %d\n", __func__, k, swath_width_luma_ub[k]); in dml32_CalculateSwathAndDETConfiguration()
521 dml_print("DML::%s: k=%0d BytePerPixDETY = %f\n", __func__, k, BytePerPixDETY[k]); in dml32_CalculateSwathAndDETConfiguration()
522 dml_print("DML::%s: k=%0d MaximumSwathHeightY = %d\n", __func__, k, MaximumSwathHeightY[k]); in dml32_CalculateSwathAndDETConfiguration()
523 dml_print("DML::%s: k=%0d RoundedUpMaxSwathSizeBytesY = %d\n", __func__, k, in dml32_CalculateSwathAndDETConfiguration()
524 RoundedUpMaxSwathSizeBytesY[k]); in dml32_CalculateSwathAndDETConfiguration()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml2_0/
H A Ddisplay_mode_core.c2698 for (dml_uint_t k = 0; k < num_active_planes; ++k) { in PixelClockAdjustmentForProgressiveToInterlaceUnit() local
2699 display_cfg->output.PixelClockBackEnd[k] = display_cfg->timing.PixelClock[k]; in PixelClockAdjustmentForProgressiveToInterlaceUnit()
2700 if (display_cfg->timing.Interlace[k] == 1 && ptoi_supported == true) { in PixelClockAdjustmentForProgressiveToInterlaceUnit()
2701 display_cfg->timing.PixelClock[k] = 2 * display_cfg->timing.PixelClock[k]; in PixelClockAdjustmentForProgressiveToInterlaceUnit()
2853 for (dml_uint_t k = 0; k < p->NumberOfActiveSurfaces; ++k) { in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport() local
2854 if (p->WritebackEnable[k] == true) { in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport()
2890 for (dml_uint_t k = 0; k < p->NumberOfActiveSurfaces; ++k) { in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport() local
2891 s->TotalPixelBW = s->TotalPixelBW + p->DPPPerSurface[k] in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport()
2892 …dthY[k] * p->BytePerPixelDETY[k] * p->VRatio[k] + p->SwathWidthC[k] * p->BytePerPixelDETC[k] * p->… in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport()
2895 for (dml_uint_t k = 0; k < p->NumberOfActiveSurfaces; ++k) { in CalculateWatermarksMALLUseAndDRAMSpeedChangeSupport() local
[all …]
/linux/drivers/md/bcache/
H A Dextents.c33 i->k = bkey_next(i->k); in sort_key_next()
35 if (i->k == i->end) in sort_key_next()
42 int64_t c = bkey_cmp(l.k, r.k); in bch_key_sort_cmp()
44 return c ? c > 0 : l.k < r.k; in bch_key_sort_cmp()
47 static bool __ptr_invalid(struct cache_set *c, const struct bkey *k) in __ptr_invalid() argument
51 for (i = 0; i < KEY_PTRS(k); i++) in __ptr_invalid()
52 if (ptr_available(c, k, i)) { in __ptr_invalid()
54 size_t bucket = PTR_BUCKET_NR(c, k, i); in __ptr_invalid()
55 size_t r = bucket_remainder(c, PTR_OFFSET(k, i)); in __ptr_invalid()
57 if (KEY_SIZE(k) + r > c->cache->sb.bucket_size || in __ptr_invalid()
[all …]
H A Dbset.c23 struct bkey *k, *next; in bch_dump_bset() local
25 for (k = i->start; k < bset_bkey_last(i); k = next) { in bch_dump_bset()
26 next = bkey_next(k); in bch_dump_bset()
29 (unsigned int) ((u64 *) k - i->d), i->keys); in bch_dump_bset()
32 b->ops->key_dump(b, k); in bch_dump_bset()
34 pr_cont("%llu:%llu\n", KEY_INODE(k), KEY_OFFSET(k)); in bch_dump_bset()
37 bkey_cmp(k, b->ops->is_extents ? in bch_dump_bset()
58 struct bkey *k; in __bch_count_data() local
61 for_each_key(b, k, &iter) in __bch_count_data()
62 ret += KEY_SIZE(k); in __bch_count_data()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/
H A Ddml2_core_dcn4_calcs.c159 for (unsigned int k = 0; k < display_cfg->num_planes; k++) { in get_stream_output_bpp() local
160 …double bpc = (double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_inde… in get_stream_output_bpp()
161 …if (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.dsc.ena… in get_stream_output_bpp()
162 …switch (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].output.out… in get_stream_output_bpp()
164 out_bpp[k] = bpc * 3; in get_stream_output_bpp()
167 out_bpp[k] = bpc * 2; in get_stream_output_bpp()
170 out_bpp[k] = bpc * 2; in get_stream_output_bpp()
174 out_bpp[k] = bpc * 1.5; in get_stream_output_bpp()
177 …} else if (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.… in get_stream_output_bpp()
178 …out_bpp[k] = (double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_inde… in get_stream_output_bpp()
[all …]
/linux/tools/lib/
H A Dbitmap.c10 unsigned int k, w = 0, lim = bits/BITS_PER_LONG; in __bitmap_weight() local
12 for (k = 0; k < lim; k++) in __bitmap_weight()
13 w += hweight_long(bitmap[k]); in __bitmap_weight()
16 w += hweight_long(bitmap[k] & BITMAP_LAST_WORD_MASK(bits)); in __bitmap_weight()
24 int k; in __bitmap_or() local
27 for (k = 0; k < nr; k++) in __bitmap_or()
28 dst[k] = bitmap1[k] | bitmap2[k]; in __bitmap_or()
63 unsigned int k; in __bitmap_and() local
67 for (k = 0; k < lim; k++) in __bitmap_and()
68 result |= (dst[k] = bitmap1[k] & bitmap2[k]); in __bitmap_and()
[all …]
H A Dlist_sort.c117 * 2:1 balanced merges. Given two pending sublists of size 2^k, they are
118 * merged to a size-2^(k+1) list as soon as we have 2^k following elements.
120 * Thus, it will avoid cache thrashing as long as 3*2^k elements can
129 * Each time we increment "count", we set one bit (bit k) and clear
130 * bits k-1 .. 0. Each time this happens (except the very first time
131 * for each bit, when count increments to 2^k), we merge two lists of
132 * size 2^k into one list of size 2^(k+1).
135 * 2^k, which is when we have 2^k elements pending in smaller lists,
136 * so it's safe to merge away two lists of size 2^k.
138 * After this happens twice, we have created two lists of size 2^(k+1),
[all …]
/linux/arch/sparc/kernel/
H A Dtermios.c33 int user_termios_to_kernel_termios(struct ktermios *k, in user_termios_to_kernel_termios() argument
37 err = get_user(k->c_iflag, &u->c_iflag); in user_termios_to_kernel_termios()
38 err |= get_user(k->c_oflag, &u->c_oflag); in user_termios_to_kernel_termios()
39 err |= get_user(k->c_cflag, &u->c_cflag); in user_termios_to_kernel_termios()
40 err |= get_user(k->c_lflag, &u->c_lflag); in user_termios_to_kernel_termios()
41 err |= get_user(k->c_line, &u->c_line); in user_termios_to_kernel_termios()
42 err |= copy_from_user(k->c_cc, u->c_cc, NCCS); in user_termios_to_kernel_termios()
43 if (k->c_lflag & ICANON) { in user_termios_to_kernel_termios()
44 err |= get_user(k->c_cc[VEOF], &u->c_cc[VEOF]); in user_termios_to_kernel_termios()
45 err |= get_user(k->c_cc[VEOL], &u->c_cc[VEOL]); in user_termios_to_kernel_termios()
[all …]
/linux/fs/btrfs/tests/
H A Dextent-map-tests.c50 * extent [0, 16K), followed by another file extent [16K, 20K), two dio reads
51 * are entering btrfs_get_extent() concurrently, t1 is reading [8K, 16K), t2 is
52 * reading [0, 8K)
57 * -> add_extent_mapping(0, 16K)
59 * ->add_extent_mapping(0, 16K)
77 /* Add [0, 16K) */ in test_case_1()
87 test_err("cannot add extent range [0, 16K)"); in test_case_1()
92 /* Add [16K, 20K) following [0, 16K) */ in test_case_1()
109 test_err("cannot add extent range [16K, 20K)"); in test_case_1()
121 /* Add [0, 8K), should return [0, 16K) instead. */ in test_case_1()
[all …]
/linux/drivers/ata/pata_parport/
H A Dfriq.c71 int h, l, k, ph; in friq_read_block_int() local
76 for (k = 0; k < count; k++) { in friq_read_block_int()
79 buf[k] = j44(l, h); in friq_read_block_int()
87 for (k = 0; k < count; k++) { in friq_read_block_int()
89 buf[k] = r0(); in friq_read_block_int()
96 for (k = 0; k < count - 2; k++) in friq_read_block_int()
97 buf[k] = r4(); in friq_read_block_int()
105 for (k = 0; k < count / 2 - 1; k++) in friq_read_block_int()
106 ((u16 *)buf)[k] = r4w(); in friq_read_block_int()
114 for (k = 0; k < count / 4 - 1; k++) in friq_read_block_int()
[all …]
H A Dfrpw.c59 int h, l, k, ph; in frpw_read_block_int() local
64 for (k = 0; k < count; k++) { in frpw_read_block_int()
67 buf[k] = j44(l, h); in frpw_read_block_int()
76 for (k = 0; k < count; k++) { in frpw_read_block_int()
78 buf[k] = r0(); in frpw_read_block_int()
86 for (k = 0; k < count; k++) in frpw_read_block_int()
87 buf[k] = r4(); in frpw_read_block_int()
94 for (k = 0; k < count - 2; k++) in frpw_read_block_int()
95 buf[k] = r4(); in frpw_read_block_int()
104 for (k = 0; k < count / 2 - 1; k++) in frpw_read_block_int()
[all …]
H A Depat.c87 int k, ph, a, b; in epat_read_block() local
94 for (k = 0; k < count; k++) { in epat_read_block()
95 if (k == count-1) in epat_read_block()
103 buf[k] = j44(a, b); in epat_read_block()
112 for (k = 0; k < count; k++) { in epat_read_block()
113 if (k == count - 1) in epat_read_block()
117 buf[k] = j53(a, b); in epat_read_block()
126 for (k = 0; k < count - 1; k++) { in epat_read_block()
128 buf[k] = r0(); in epat_read_block()
138 for (k = 0; k < count - 1; k++) in epat_read_block()
[all …]
H A Depia.c129 int k, ph, a, b; in epia_read_block() local
135 for (k = 0; k < count; k++) { in epia_read_block()
138 buf[k] = j44(a, b); in epia_read_block()
147 for (k = 0; k < count; k++) { in epia_read_block()
150 buf[k] = j53(a, b); in epia_read_block()
158 for (k = 0; k < count; k++) { in epia_read_block()
160 buf[k] = r0(); in epia_read_block()
169 for (k = 0; k < count; k++) in epia_read_block()
170 buf[k] = r4(); in epia_read_block()
177 for (k = 0; k < count / 2; k++) in epia_read_block()
[all …]
H A Dkbic.c128 int k, a, b; in kbic_read_block() local
133 for (k = 0; k < count / 2; k++) { in kbic_read_block()
138 buf[2 * k] = j44(a, b); in kbic_read_block()
143 buf[2 * k + 1] = j44(a, b); in kbic_read_block()
149 for (k = 0; k < count / 4; k++) { in kbic_read_block()
153 buf[4 * k] = j53(r12w()); in kbic_read_block()
155 buf[4 * k + 1] = j53(r12w()); in kbic_read_block()
157 buf[4 * k + 3] = j53(r12w()); in kbic_read_block()
159 buf[4 * k + 2] = j53(r12w()); in kbic_read_block()
165 for (k = 0; k < count / 2; k++) { in kbic_read_block()
[all …]
H A Ddstr.c117 int k, a, b; in dstr_read_block() local
128 for (k = 0; k < count; k++) { in dstr_read_block()
131 buf[k] = j44(a, b); in dstr_read_block()
136 for (k = 0; k < count; k++) { in dstr_read_block()
138 buf[k] = r0(); in dstr_read_block()
145 for (k = 0; k < count; k++) in dstr_read_block()
146 buf[k] = r4(); in dstr_read_block()
151 for (k = 0; k < count / 2; k++) in dstr_read_block()
152 ((u16 *)buf)[k] = r4w(); in dstr_read_block()
157 for (k = 0; k < count / 4; k++) in dstr_read_block()
[all …]
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dhashmap.c14 static size_t hash_fn(long k, void *ctx) in hash_fn() argument
16 return k; in hash_fn()
56 long oldk, k = i; in test_hashmap_generic() local
59 err = hashmap__update(map, k, v, &oldk, &oldv); in test_hashmap_generic()
65 err = hashmap__add(map, k, v); in test_hashmap_generic()
67 err = hashmap__set(map, k, v, &oldk, &oldv); in test_hashmap_generic()
69 "unexpected k/v: %ld=%ld\n", oldk, oldv)) in test_hashmap_generic()
73 if (CHECK(err, "elem_add", "failed to add k/v %ld = %ld: %d\n", k, v, err)) in test_hashmap_generic()
76 if (CHECK(!hashmap__find(map, k, &oldv), "elem_find", in test_hashmap_generic()
77 "failed to find key %ld\n", k)) in test_hashmap_generic()
[all …]
/linux/lib/crypto/
H A Dsm3.c18 static const u32 ____cacheline_aligned K[64] = { variable
88 R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4)); in sm3_transform()
89 R1(d, a, b, c, h, e, f, g, K[1], I(1), I(5)); in sm3_transform()
90 R1(c, d, a, b, g, h, e, f, K[2], I(2), I(6)); in sm3_transform()
91 R1(b, c, d, a, f, g, h, e, K[3], I(3), I(7)); in sm3_transform()
92 R1(a, b, c, d, e, f, g, h, K[4], W1(4), I(8)); in sm3_transform()
93 R1(d, a, b, c, h, e, f, g, K[5], W1(5), I(9)); in sm3_transform()
94 R1(c, d, a, b, g, h, e, f, K[6], W1(6), I(10)); in sm3_transform()
95 R1(b, c, d, a, f, g, h, e, K[7], W1(7), I(11)); in sm3_transform()
96 R1(a, b, c, d, e, f, g, h, K[8], W1(8), I(12)); in sm3_transform()
[all …]
/linux/rust/kernel/
H A Drbtree.rs171 pub struct RBTree<K, V> {
173 _p: PhantomData<Node<K, V>>,
177 // fields, so we use the same Send condition as would be used for a struct with K and V fields.
178 unsafe impl<K: Send, V: Send> Send for RBTree<K, V> {}
181 // fields, so we use the same Sync condition as would be used for a struct with K and V fields.
182 unsafe impl<K: Sync, V: Sync> Sync for RBTree<K, V> {}
184 impl<K, V> RBTree<K,
201 iter(&self) -> Iter<'_, K, V> iter() argument
216 iter_mut(&mut self) -> IterMut<'_, K, V> iter_mut() argument
246 cursor_front_mut(&mut self) -> Option<CursorMut<'_, K, V>> cursor_front_mut() argument
261 cursor_front(&self) -> Option<Cursor<'_, K, V>> cursor_front() argument
276 cursor_back_mut(&mut self) -> Option<CursorMut<'_, K, V>> cursor_back_mut() argument
291 cursor_back(&self) -> Option<Cursor<'_, K, V>> cursor_back() argument
318 try_create_and_insert( &mut self, key: K, value: V, flags: Flags, ) -> Result<Option<RBTreeNode<K, V>>> try_create_and_insert() argument
321 try_create_and_insert( &mut self, key: K, value: V, flags: Flags, ) -> Result<Option<RBTreeNode<K, V>>> try_create_and_insert() argument
331 insert(&mut self, node: RBTreeNode<K, V>) -> Option<RBTreeNode<K, V>> insert() argument
341 raw_entry(&mut self, key: &K) -> RawEntry<'_, K, V> raw_entry() argument
395 entry(&mut self, key: K) -> Entry<'_, K, V> entry() argument
403 find_mut(&mut self, key: &K) -> Option<OccupiedEntry<'_, K, V>> find_mut() argument
438 remove_node(&mut self, key: &K) -> Option<RBTreeNode<K, V>> remove_node() argument
454 cursor_lower_bound_mut(&mut self, key: &K) -> Option<CursorMut<'_, K, V>> where K: Ord, cursor_lower_bound_mut() argument
475 cursor_lower_bound(&self, key: &K) -> Option<Cursor<'_, K, V>> where K: Ord, cursor_lower_bound() argument
826 current(&self) -> (&K, &V) current() argument
837 to_key_value<'b>(node: NonNull<bindings::rb_node>) -> (&'b K, &'b V) to_key_value() argument
843 let k = unsafe { &(*this).key }; to_key_value() localVariable
851 peek_prev(&self) -> Option<(&K, &V)> peek_prev() argument
856 peek_next(&self) -> Option<(&K, &V)> peek_next() argument
860 peek(&self, direction: Direction) -> Option<(&K, &V)> peek() argument
895 current(&self) -> (&K, &V) current() argument
903 current_mut(&mut self) -> (&K, &mut V) current_mut() argument
915 remove_current(self) -> (Option<Self>, RBTreeNode<K, V>) remove_current() argument
939 remove_prev(&mut self) -> Option<RBTreeNode<K, V>> remove_prev() argument
944 remove_next(&mut self) -> Option<RBTreeNode<K, V>> remove_next() argument
948 remove_neighbor(&mut self, direction: Direction) -> Option<RBTreeNode<K, V>> remove_neighbor() argument
984 peek_prev(&self) -> Option<(&K, &V)> peek_prev() argument
989 peek_next(&self) -> Option<(&K, &V)> peek_next() argument
993 peek(&self, direction: Direction) -> Option<(&K, &V)> peek() argument
1003 peek_prev_mut(&mut self) -> Option<(&K, &mut V)> peek_prev_mut() argument
1008 peek_next_mut(&mut self) -> Option<(&K, &mut V)> peek_next_mut() argument
1012 peek_mut(&mut self, direction: Direction) -> Option<(&K, &mut V)> peek_mut() argument
1037 to_key_value<'b>(node: NonNull<bindings::rb_node>) -> (&'b K, &'b V) to_key_value() argument
1048 to_key_value_mut<'b>(node: NonNull<bindings::rb_node>) -> (&'b K, &'b mut V) to_key_value_mut() argument
1059 to_key_value_raw<'b>(node: NonNull<bindings::rb_node>) -> (&'b K, *mut V) to_key_value_raw() argument
1065 let k = unsafe { &(*this).key }; to_key_value_raw() localVariable
1193 new(flags: Flags) -> Result<RBTreeNodeReservation<K, V>> new() argument
1211 into_node(self, key: K, value: V) -> RBTreeNode<K, V> into_node() argument
1235 new(key: K, value: V, flags: Flags) -> Result<RBTreeNode<K, V>> new() argument
1240 to_key_value(self) -> (K, V) to_key_value() argument
1263 into_reservation(self) -> RBTreeNodeReservation<K, V> into_reservation() argument
1315 insert(self, node: RBTreeNode<K, V>) -> &'a mut V insert() argument
1337 insert(self, value: V, reservation: RBTreeNodeReservation<K, V>) -> &'a mut V insert() argument
1380 remove_node(self) -> RBTreeNode<K, V> remove_node() argument
1404 replace(self, node: RBTreeNode<K, V>) -> RBTreeNode<K, V> replace() argument
[all...]
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_jhash.h39 const unsigned char *k = key; in jhash() local
44 a += *(volatile u32 *)(k); in jhash()
45 b += *(volatile u32 *)(k + 4); in jhash()
46 c += *(volatile u32 *)(k + 8); in jhash()
49 k += 12; in jhash()
52 case 12: c += (u32)k[11]<<24; in jhash()
53 case 11: c += (u32)k[10]<<16; in jhash()
54 case 10: c += (u32)k[9]<<8; in jhash()
55 case 9: c += k[8]; in jhash()
56 case 8: b += (u32)k[7]<<24; in jhash()
[all …]
/linux/net/ceph/
H A Dceph_hash.c25 const unsigned char *k = (const unsigned char *)str; in ceph_str_hash_rjenkins() local
37 a = a + (k[0] + ((__u32)k[1] << 8) + ((__u32)k[2] << 16) + in ceph_str_hash_rjenkins()
38 ((__u32)k[3] << 24)); in ceph_str_hash_rjenkins()
39 b = b + (k[4] + ((__u32)k[5] << 8) + ((__u32)k[6] << 16) + in ceph_str_hash_rjenkins()
40 ((__u32)k[7] << 24)); in ceph_str_hash_rjenkins()
41 c = c + (k[8] + ((__u32)k[9] << 8) + ((__u32)k[10] << 16) + in ceph_str_hash_rjenkins()
42 ((__u32)k[11] << 24)); in ceph_str_hash_rjenkins()
44 k = k + 12; in ceph_str_hash_rjenkins()
52 c = c + ((__u32)k[10] << 24); in ceph_str_hash_rjenkins()
55 c = c + ((__u32)k[9] << 16); in ceph_str_hash_rjenkins()
[all …]
/linux/lib/
H A Dklist.c74 * @k: The klist we're initializing.
84 void klist_init(struct klist *k, void (*get)(struct klist_node *), in klist_init() argument
87 INIT_LIST_HEAD(&k->k_list); in klist_init()
88 spin_lock_init(&k->k_lock); in klist_init()
89 k->get = get; in klist_init()
90 k->put = put; in klist_init()
94 static void add_head(struct klist *k, struct klist_node *n) in add_head() argument
96 spin_lock(&k->k_lock); in add_head()
97 list_add(&n->n_node, &k->k_list); in add_head()
98 spin_unlock(&k->k_lock); in add_head()
[all …]

12345678910>>...111