| /linux/tools/perf/pmu-events/ |
| H A D | metric.py | 122 rhs: Union[int, float, Expression]): 125 self.rhs = _Constify(rhs) 130 rhs: bool = False) -> str: 154 if rhs and _PRECEDENCE.get(self.operator, -1) == _PRECEDENCE.get( 169 rhs = self.rhs.Simplify() 170 if isinstance(lhs, Constant) and isinstance(rhs, Constant): 171 return Constant(ast.literal_eval(lhs + self.operator + rhs)) 175 return rhs 180 not isinstance(rhs, Event) or 'slots' not in rhs.name.lower()): 184 return rhs [all …]
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | evlist.c | 8 int arch_evlist__cmp(const struct evsel *lhs, const struct evsel *rhs) in arch_evlist__cmp() argument 51 (arch_evsel__must_be_in_group(lhs) || arch_evsel__must_be_in_group(rhs))) { in arch_evlist__cmp() 55 if (arch_is_topdown_slots(rhs)) in arch_evlist__cmp() 64 if (lhs->core.leader != rhs->core.leader) { in arch_evlist__cmp() 66 bool rhs_topdown = arch_is_topdown_metrics(rhs); in arch_evlist__cmp() 76 if (lhs->retire_lat && !rhs->retire_lat) in arch_evlist__cmp() 78 if (!lhs->retire_lat && rhs->retire_lat) in arch_evlist__cmp() 82 return lhs->core.idx - rhs->core.idx; in arch_evlist__cmp()
|
| /linux/rust/kernel/sync/atomic/ |
| H A D | predefine.rs | 16 fn rhs_into_delta(rhs: i32) -> i32 { in rhs_into_delta() 17 rhs in rhs_into_delta() 29 fn rhs_into_delta(rhs: i64) -> i64 { in rhs_into_delta() 30 rhs in rhs_into_delta() 58 fn rhs_into_delta(rhs: isize) -> isize_atomic_repr { in rhs_into_delta() 59 rhs as isize_atomic_repr in rhs_into_delta() 71 fn rhs_into_delta(rhs: u32) -> i32 { in rhs_into_delta() 72 rhs as i32 in rhs_into_delta() 84 fn rhs_into_delta(rhs: u64) -> i64 { in rhs_into_delta() 85 rhs as i64 in rhs_into_delta() [all …]
|
| /linux/rust/kernel/num/ |
| H A D | bounded.rs | 593 fn add(self, rhs: Bounded<T, M>) -> Self::Output { in add() 594 self.get() + rhs.get() in add() 605 fn bitand(self, rhs: Bounded<T, M>) -> Self::Output { in bitand() 606 self.get() & rhs.get() in bitand() 617 fn bitor(self, rhs: Bounded<T, M>) -> Self::Output { in bitor() 618 self.get() | rhs.get() in bitor() 629 fn bitxor(self, rhs: Bounded<T, M>) -> Self::Output { in bitxor() 630 self.get() ^ rhs.get() in bitxor() 641 fn div(self, rhs: Bounded<T, M>) -> Self::Output { in div() 642 self.get() / rhs.get() in div() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | gen-hyprel.c | 180 #define assert_op(lhs, rhs, fmt, op) \ argument 183 typeof(rhs) _rhs = (rhs); \ 186 fatal_error("assertion " #lhs " " #op " " #rhs \ 192 #define assert_eq(lhs, rhs, fmt) assert_op(lhs, rhs, fmt, ==) argument 193 #define assert_ne(lhs, rhs, fmt) assert_op(lhs, rhs, fmt, !=) argument 194 #define assert_lt(lhs, rhs, fmt) assert_op(lhs, rhs, fmt, <) argument 195 #define assert_ge(lhs, rhs, fmt) assert_op(lhs, rhs, fmt, >=) argument
|
| /linux/lib/ |
| H A D | ubsan.c | 242 void *rhs, char op) in handle_overflow() argument 257 val_to_string(rhs_val_str, sizeof(rhs_val_str), type, rhs); in handle_overflow() 268 void *lhs, void *rhs) in __ubsan_handle_add_overflow() argument 271 handle_overflow(data, lhs, rhs, '+'); in __ubsan_handle_add_overflow() 276 void *lhs, void *rhs) in __ubsan_handle_sub_overflow() argument 278 handle_overflow(data, lhs, rhs, '-'); in __ubsan_handle_sub_overflow() 283 void *lhs, void *rhs) in __ubsan_handle_mul_overflow() argument 285 handle_overflow(data, lhs, rhs, '*'); in __ubsan_handle_mul_overflow() 333 void __ubsan_handle_divrem_overflow(void *_data, void *lhs, void *rhs) in __ubsan_handle_divrem_overflow() argument 345 if (type_is_signed(data->type) && get_signed_val(data->type, rhs) == -1) in __ubsan_handle_divrem_overflow() [all …]
|
| H A D | percpu_counter.c | 292 int __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch) in __percpu_counter_compare() argument 298 if (abs(count - rhs) > (batch * num_online_cpus())) { in __percpu_counter_compare() 299 if (count > rhs) in __percpu_counter_compare() 306 if (count > rhs) in __percpu_counter_compare() 308 else if (count < rhs) in __percpu_counter_compare()
|
| /linux/include/asm-generic/ |
| H A D | word-at-a-time.h | 18 static inline long prep_zero_mask(unsigned long val, unsigned long rhs, const struct word_at_a_time… in prep_zero_mask() argument 21 return ~(mask | rhs); in prep_zero_mask() 44 unsigned long rhs = val | c->low_bits; in has_zero() local 45 *data = rhs; in has_zero() 46 return (val + c->high_bits) & ~rhs; in has_zero()
|
| /linux/include/linux/ |
| H A D | percpu_counter.h | 59 int __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch); 64 static inline int percpu_counter_compare(struct percpu_counter *fbc, s64 rhs) in percpu_counter_compare() argument 66 return __percpu_counter_compare(fbc, rhs, percpu_counter_batch); in percpu_counter_compare() 171 static inline int percpu_counter_compare(struct percpu_counter *fbc, s64 rhs) in percpu_counter_compare() argument 173 if (fbc->count > rhs) in percpu_counter_compare() 175 else if (fbc->count < rhs) in percpu_counter_compare() 182 __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch) in __percpu_counter_compare() argument 184 return percpu_counter_compare(fbc, rhs); in __percpu_counter_compare()
|
| H A D | iosys-map.h | 215 const struct iosys_map *rhs) in iosys_map_is_equal() argument 217 if (lhs->is_iomem != rhs->is_iomem) in iosys_map_is_equal() 220 return lhs->vaddr_iomem == rhs->vaddr_iomem; in iosys_map_is_equal() 222 return lhs->vaddr == rhs->vaddr; in iosys_map_is_equal()
|
| /linux/tools/testing/selftests/powerpc/primitives/ |
| H A D | word-at-a-time.h | 21 static inline long prep_zero_mask(unsigned long val, unsigned long rhs, const struct word_at_a_time… in prep_zero_mask() argument 24 return ~(mask | rhs); in prep_zero_mask() 39 unsigned long rhs = val | c->low_bits; in has_zero() local 40 *data = rhs; in has_zero() 41 return (val + c->high_bits) & ~rhs; in has_zero()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | word-at-a-time.h | 21 static inline long prep_zero_mask(unsigned long val, unsigned long rhs, const struct word_at_a_time… in prep_zero_mask() argument 24 return ~(mask | rhs); in prep_zero_mask() 39 unsigned long rhs = val | c->low_bits; in has_zero() local 40 *data = rhs; in has_zero() 41 return (val + c->high_bits) & ~rhs; in has_zero()
|
| /linux/tools/perf/scripts/python/ |
| H A D | compaction-times.py | 61 def __add__(self, rhs): argument 62 self.aval += rhs.aval 63 self.bval += rhs.bval 76 def __add__(self, rhs): argument 77 self.ns += rhs.ns 78 self.migrated += rhs.migrated 79 self.fscan += rhs.fscan 80 self.mscan += rhs.mscan 167 def __add__(self, rhs): argument 168 self.ns += rhs.ns [all …]
|
| /linux/drivers/md/ |
| H A D | dm-bio-prison-v2.c | 89 struct dm_cell_key_v2 *rhs) in cmp_keys() argument 91 if (lhs->virtual < rhs->virtual) in cmp_keys() 94 if (lhs->virtual > rhs->virtual) in cmp_keys() 97 if (lhs->dev < rhs->dev) in cmp_keys() 100 if (lhs->dev > rhs->dev) in cmp_keys() 103 if (lhs->block_end <= rhs->block_begin) in cmp_keys() 106 if (lhs->block_begin >= rhs->block_end) in cmp_keys()
|
| /linux/rust/kernel/ |
| H A D | alloc.rs | 50 fn bitor(self, rhs: Self) -> Self::Output { in bitor() 51 Self(self.0 | rhs.0) in bitor() 57 fn bitand(self, rhs: Self) -> Self::Output { in bitand() 58 Self(self.0 & rhs.0) in bitand()
|
| H A D | dma.rs | 205 fn bitor(self, rhs: Self) -> Self::Output { in bitor() 206 Self(self.0 | rhs.0) in bitor() 212 fn bitand(self, rhs: Self) -> Self::Output { in bitand() 213 Self(self.0 & rhs.0) in bitand()
|
| /linux/rust/kernel/io/ |
| H A D | resource.rs | 197 fn bitor(self, rhs: Self) -> Self::Output { in bitor() 198 Self(self.0 | rhs.0) in bitor() 204 fn bitand(self, rhs: Self) -> Self::Output { in bitand() 205 Self(self.0 & rhs.0) in bitand()
|
| /linux/scripts/gcc-plugins/ |
| H A D | latent_entropy_plugin.c | 301 static enum tree_code get_op(tree *rhs) in get_op() argument 314 if (rhs) { in get_op() 329 if (rhs) in get_op() 330 *rhs = build_int_cstu(long_unsigned_type_node, random_const); in get_op() 344 tree rhs; in perturb_local_entropy() local 347 op = get_op(&rhs); in perturb_local_entropy() 348 assign = create_assign(op, local_entropy, local_entropy, rhs); in perturb_local_entropy()
|
| /linux/include/linux/ceph/ |
| H A D | msgr.h | 91 const struct ceph_entity_addr *rhs) in ceph_addr_equal_no_type() argument 93 return !memcmp(&lhs->in_addr, &rhs->in_addr, sizeof(lhs->in_addr)) && in ceph_addr_equal_no_type() 94 lhs->nonce == rhs->nonce; in ceph_addr_equal_no_type()
|
| /linux/arch/riscv/kernel/tests/module_test/ |
| H A D | test_module_linking_main.c | 29 #define CHECK_EQ(lhs, rhs) KUNIT_ASSERT_EQ(test, lhs, rhs) argument
|
| /linux/tools/lib/perf/ |
| H A D | cpumap.c | 325 bool perf_cpu_map__equal(const struct perf_cpu_map *lhs, const struct perf_cpu_map *rhs) in perf_cpu_map__equal() argument 329 if (lhs == rhs) in perf_cpu_map__equal() 332 if (!lhs || !rhs) in perf_cpu_map__equal() 336 if (nr != __perf_cpu_map__nr(rhs)) in perf_cpu_map__equal() 340 if (__perf_cpu_map__cpu(lhs, idx).cpu != __perf_cpu_map__cpu(rhs, idx).cpu) in perf_cpu_map__equal()
|
| /linux/tools/testing/selftests/cgroup/ |
| H A D | test_cpu.c | 108 timespec_sub(const struct timespec *lhs, const struct timespec *rhs) in timespec_sub() argument 116 if (lhs->tv_sec < rhs->tv_sec) in timespec_sub() 119 ret.tv_sec = lhs->tv_sec - rhs->tv_sec; in timespec_sub() 121 if (lhs->tv_nsec < rhs->tv_nsec) { in timespec_sub() 126 ret.tv_nsec = NSEC_PER_SEC - rhs->tv_nsec + lhs->tv_nsec; in timespec_sub() 128 ret.tv_nsec = lhs->tv_nsec - rhs->tv_nsec; in timespec_sub()
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_metafile.c | 85 int64_t rhs) in xfs_metafile_resv_can_cover() argument 93 if (mp->m_metafile_resv_avail >= rhs) in xfs_metafile_resv_can_cover() 101 rhs - mp->m_metafile_resv_avail, 2048) >= 0; in xfs_metafile_resv_can_cover()
|
| /linux/rust/macros/ |
| H A D | fmt.rs | 55 let (lhs, rhs) = (|| { in fmt() 70 args.extend(quote_spanned!(first_span => #lhs #adapter(&#rhs))); in fmt()
|
| /linux/lib/zstd/common/ |
| H A D | compiler.h | 234 ptrdiff_t ZSTD_wrappedPtrDiff(unsigned char const* lhs, unsigned char const* rhs) in ZSTD_wrappedPtrDiff() argument 236 return lhs - rhs; in ZSTD_wrappedPtrDiff()
|