| /linux/drivers/input/ |
| H A D | ff-memless.c | 110 static void ml_schedule_timer(struct ml_device *ml) in ml_schedule_timer() argument 123 state = &ml->states[i]; in ml_schedule_timer() 140 timer_delete(&ml->timer); in ml_schedule_timer() 143 mod_timer(&ml->timer, earliest); in ml_schedule_timer() 324 static int ml_get_combo_effect(struct ml_device *ml, in ml_get_combo_effect() argument 339 state = &ml->states[i]; in ml_get_combo_effect() 353 effect_type = get_compatible_type(ml->dev->ff, effect->type); in ml_get_combo_effect() 381 ml_combine_effects(combo_effect, state, ml->gain); in ml_get_combo_effect() 388 static void ml_play_effects(struct ml_device *ml) in ml_play_effects() argument 395 while (ml_get_combo_effect(ml, handled_bm, &effect)) in ml_play_effects() [all …]
|
| /linux/fs/ocfs2/dlm/ |
| H A D | dlmast.c | 52 if (lock->ml.highest_blocked == LKM_IVMODE) in dlm_should_cancel_bast() 54 BUG_ON(lock->ml.highest_blocked == LKM_NLMODE); in dlm_should_cancel_bast() 61 if (lock->ml.type == LKM_EXMODE) in dlm_should_cancel_bast() 64 else if (lock->ml.type == LKM_NLMODE) in dlm_should_cancel_bast() 67 else if (lock->ml.highest_blocked != LKM_EXMODE) in dlm_should_cancel_bast() 89 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 90 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 91 lock->ast_pending, lock->ml.type); in __dlm_queue_ast() 97 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 98 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie))); in __dlm_queue_ast() [all …]
|
| H A D | dlmthread.c | 365 if (target->ml.convert_type == LKM_IVMODE) { in dlm_shuffle_lists() 373 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 374 target->ml.convert_type)) { in dlm_shuffle_lists() 377 if (lock->ml.highest_blocked == LKM_IVMODE) { in dlm_shuffle_lists() 382 if (lock->ml.highest_blocked < target->ml.convert_type) in dlm_shuffle_lists() 383 lock->ml.highest_blocked = in dlm_shuffle_lists() 384 target->ml.convert_type; in dlm_shuffle_lists() 391 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 392 target->ml.convert_type)) { in dlm_shuffle_lists() 394 if (lock->ml.highest_blocked == LKM_IVMODE) { in dlm_shuffle_lists() [all …]
|
| H A D | dlmconvert.c | 114 lock->ml.type, lock->ml.convert_type, type); in __dlmconvert_master() 119 if (lock->ml.convert_type != LKM_IVMODE) { in __dlmconvert_master() 135 switch (lock->ml.type) { in __dlmconvert_master() 139 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 149 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 155 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 165 if (type <= lock->ml.type) in __dlmconvert_master() 173 if (!dlm_lock_compatible(tmplock->ml.type, type)) in __dlmconvert_master() 178 if (!dlm_lock_compatible(tmplock->ml.type, type)) in __dlmconvert_master() 181 if (!dlm_lock_compatible(tmplock->ml.convert_type, type)) in __dlmconvert_master() [all …]
|
| H A D | dlmrecovery.c | 1047 if (lock->ml.node == dead_node) { in dlm_move_reco_locks_to_list() 1122 struct_size(mres, ml, mres->num_locks), in dlm_send_mig_lockres_msg() 1181 if (lock->ml.type != LKM_EXMODE && lock->ml.type != LKM_PRMODE) in dlm_prepare_lvb_for_migration() 1195 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1196 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1198 lock->ml.node); in dlm_prepare_lvb_for_migration() 1208 struct dlm_migratable_lock *ml; in dlm_add_lock_to_array() local 1211 ml = &(mres->ml[lock_num]); in dlm_add_lock_to_array() 1212 ml->cookie = lock->ml.cookie; in dlm_add_lock_to_array() 1213 ml->type = lock->ml.type; in dlm_add_lock_to_array() [all …]
|
| H A D | dlmlock.c | 79 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 84 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 86 if (!dlm_lock_compatible(tmplock->ml.convert_type, in dlm_can_grant_new_lock() 87 lock->ml.type)) in dlm_can_grant_new_lock() 108 mlog(0, "type=%d\n", lock->ml.type); in dlmlock_master() 115 lock->ml.node != dlm->node_num) { in dlmlock_master() 144 lock->ml.node); in dlmlock_master() 155 lock->ml.node); in dlmlock_master() 205 lock->ml.type, res->lockname.len, in dlmlock_remote() 293 create.requested_type = lock->ml.type; in dlm_send_remote_lock_request() [all …]
|
| H A D | dlmunlock.c | 218 lock->ml.convert_type = LKM_IVMODE; in dlmunlock_common() 227 BUG_ON(lock->ml.convert_type != LKM_IVMODE); in dlmunlock_common() 229 BUG_ON(lock->ml.convert_type == LKM_IVMODE); in dlmunlock_common() 250 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 251 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 277 lock->ml.convert_type = LKM_IVMODE; in dlm_commit_pending_cancel() 336 unlock.cookie = lock->ml.cookie; in dlm_send_remote_unlock_request() 464 if (iter->ml.cookie == unlock->cookie && in dlm_unlock_lock_handler() 465 iter->ml.node == unlock->node_idx) { in dlm_unlock_lock_handler() 485 lock->ml.type != LKM_EXMODE) in dlm_unlock_lock_handler() [all …]
|
| H A D | dlmcommon.h | 228 struct dlm_mig_lockres_priv ml; member 337 struct dlm_migratable_lock ml; member 562 struct dlm_migratable_lock ml[]; // 16 bytes each, begins at byte 112 member
|
| H A D | dlmmaster.c | 2497 if (lock->ml.node != dlm->node_num) { in dlm_is_lockres_migratable() 2501 cookie = be64_to_cpu(lock->ml.cookie); in dlm_is_lockres_migratable() 2924 if (lock->ml.node != dlm->node_num) { in dlm_remove_nonlocal_locks() 2926 lock->ml.node); in dlm_remove_nonlocal_locks() 2933 lock->ml.node); in dlm_remove_nonlocal_locks() 2981 if (lock->ml.node == dlm->node_num) in dlm_pick_migration_target() 2983 if (test_bit(lock->ml.node, dlm->exit_domain_map)) in dlm_pick_migration_target() 2985 nodenum = lock->ml.node; in dlm_pick_migration_target()
|
| /linux/arch/arm/vfp/ |
| H A D | vfp.h | 44 static inline void add128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in add128() argument 51 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in add128() 57 static inline void sub128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in sub128() argument 64 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in sub128() 72 u32 nh, nl, mh, ml; in mul64to128() local 76 ml = m; in mul64to128() 77 rl = (u64)nl * ml; in mul64to128() 80 rma = (u64)nh * ml; in mul64to128() 112 u64 mh, ml, remh, reml, termh, terml, z; in vfp_estimate_div128to64() local 126 ml = m << 32; in vfp_estimate_div128to64() [all …]
|
| /linux/arch/sh/kernel/cpu/sh2a/ |
| H A D | fpu.c | 154 unsigned long long ml, int n) in rshift64() argument 158 return (mh << (64 - n)) | (ml >> n); in rshift64() 164 unsigned long long mh, ml, nh, nl; in denormal_muld() local 175 mult64(ix, iy, &mh, &ml); in denormal_muld() 177 nl = ml; in denormal_muld() 188 ix = (rshift64(mh, ml, w - 52) & 0x000fffffffffffffLL) in denormal_muld() 191 ix = rshift64(mh, ml, w - 51 - exp) & 0x000fffffffffffffLL; in denormal_muld()
|
| /linux/lib/zstd/compress/ |
| H A D | zstd_lazy.c | 530 size_t ZSTD_dedicatedDictSearch_lazy_search(size_t* offsetPtr, size_t ml, U32 nbAttempts, in ZSTD_dedicatedDictSearch_lazy_search() argument 563 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 576 if (currentMl > ml) { in ZSTD_dedicatedDictSearch_lazy_search() 577 ml = currentMl; in ZSTD_dedicatedDictSearch_lazy_search() 581 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 613 if (currentMl > ml) { in ZSTD_dedicatedDictSearch_lazy_search() 614 ml = currentMl; in ZSTD_dedicatedDictSearch_lazy_search() 620 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 691 size_t ml=4-1; in ZSTD_HcFindBestMatch() local 715 if (MEM_read32(match + ml - 3) == MEM_read32(ip + ml - 3)) /* potentially better */ in ZSTD_HcFindBestMatch() [all …]
|
| H A D | zstd_opt.c | 189 { unsigned ml; in ZSTD_rescaleFreqs() local 193 for (ml=0; ml<=MaxML; ml++) { in ZSTD_rescaleFreqs() 195 U32 const bitCost = FSE_getMaxNbBits(mlstate.symbolTT, ml); in ZSTD_rescaleFreqs() 197 … optPtr->matchLengthFreq[ml] = bitCost ? 1 << (scaleLog-bitCost) : 1 /*minimum to calculate cost*/; in ZSTD_rescaleFreqs() 198 optPtr->matchLengthSum += optPtr->matchLengthFreq[ml]; in ZSTD_rescaleFreqs() 234 { unsigned ml; in ZSTD_rescaleFreqs() local 235 for (ml=0; ml<=MaxML; ml++) in ZSTD_rescaleFreqs() 236 optPtr->matchLengthFreq[ml] = 1; in ZSTD_rescaleFreqs()
|
| /linux/net/packet/ |
| H A D | af_packet.c | 3666 struct packet_mclist *ml; in packet_dev_mclist_delete() 3668 while ((ml = *mlp) != NULL) { in packet_dev_mclist_delete() 3669 if (ml->ifindex == dev->ifindex) { in packet_dev_mclist_delete() 3670 list_add(&ml->remove_list, list); in packet_dev_mclist_delete() 3671 *mlp = ml->next; in packet_dev_mclist_delete() 3673 mlp = &ml->next; in packet_dev_mclist_delete() 3680 struct packet_mclist *ml, *i; in packet_mc_add() 3701 for (ml = po->mclist; ml; ml in packet_mc_add() 3665 struct packet_mclist *ml; packet_dev_mclist_delete() local 3679 struct packet_mclist *ml, *i; packet_mc_add() local 3734 struct packet_mclist *ml, **mlp; packet_mc_drop() local 3761 struct packet_mclist *ml; packet_flush_mclist() local 4185 struct packet_mclist *ml, *tmp; packet_notifier() local [all...] |
| /linux/arch/sparc/kernel/ |
| H A D | chmc.c | 404 const void *ml; in jbusmc_probe() local 453 ml = of_get_property(op->dev.of_node, "memory-layout", &p->layout_len); in jbusmc_probe() 454 if (!ml) { in jbusmc_probe() 463 memcpy(&p->layout, ml, p->layout_len); in jbusmc_probe()
|
| /linux/drivers/s390/crypto/ |
| H A D | ap_card.c | 244 ac->maxmsgsize = hwinfo.ml > 0 ? in ap_card_create() 245 hwinfo.ml * AP_TAPQ_ML_FIELD_CHUNK_SIZE : AP_DEFAULT_MAX_MSG_SIZE; in ap_card_create()
|
| /linux/net/wireless/ |
| H A D | mlme.c | 400 const struct element *ml; in cfg80211_mlme_check_mlo() local 412 ml = cfg80211_find_ext_elem(WLAN_EID_EXT_EHT_MULTI_LINK, in cfg80211_mlme_check_mlo() 414 if (!ml) { in cfg80211_mlme_check_mlo() 420 if (!ieee80211_mle_type_ok(ml->data + 1, in cfg80211_mlme_check_mlo() 422 ml->datalen - 1)) { in cfg80211_mlme_check_mlo() 428 mles[i] = (const void *)(ml->data + 1); in cfg80211_mlme_check_mlo()
|
| /linux/init/ |
| H A D | initramfs.c | 380 int ml = maybe_link(); in do_name() local 381 if (ml >= 0) { in do_name() 383 if (ml != 1) in do_name()
|
| /linux/drivers/crypto/ |
| H A D | sa2ul.c | 1045 size_t ml, pl; in sa_aes_dma_in_callback() local 1053 &ml); in sa_aes_dma_in_callback() 1090 size_t pl, ml, split_size; in sa_run() local 1274 mdptr = (u32 *)dmaengine_desc_get_metadata_ptr(tx_out, &pl, &ml); in sa_run() 1280 ml = sa_ctx->cmdl_size + (SA_PSDATA_CTX_WORDS * sizeof(u32)); in sa_run() 1360 size_t ml, pl; in sa_sha_dma_in_callback() local 1369 mdptr = (__be32 *)dmaengine_desc_get_metadata_ptr(rxd->tx_in, &pl, &ml); in sa_sha_dma_in_callback() 1668 size_t pl, ml; in sa_aead_dma_in_callback() local 1679 mdptr = (u32 *)dmaengine_desc_get_metadata_ptr(rxd->tx_in, &pl, &ml); in sa_aead_dma_in_callback()
|
| /linux/net/mac80211/ |
| H A D | parse.c | 866 const struct ieee80211_multi_link_elem *ml = elems->ml_basic; in ieee80211_mle_get_sta_prof() local 870 for_each_mle_subelement(sub, (u8 *)ml, ml_len) { in ieee80211_mle_get_sta_prof() 894 (u8 *)ml, ml_len, in ieee80211_mle_get_sta_prof()
|
| /linux/drivers/net/wireless/ath/ath12k/ |
| H A D | wmi.c | 1096 if (!restart && arg->ml.enabled) { in ath12k_wmi_vdev_start() 1098 TLV_HDR_SIZE + (arg->ml.num_partner_links * in ath12k_wmi_vdev_start() 1166 ml_params->flags = le32_encode_bits(arg->ml.enabled, in ath12k_wmi_vdev_start() 1168 le32_encode_bits(arg->ml.assoc_link, in ath12k_wmi_vdev_start() 1170 le32_encode_bits(arg->ml.mcast_link, in ath12k_wmi_vdev_start() 1172 le32_encode_bits(arg->ml.link_add, in ath12k_wmi_vdev_start() 1182 arg->ml.num_partner_links * in ath12k_wmi_vdev_start() 1188 for (i = 0; i < arg->ml.num_partner_links; i++) { in ath12k_wmi_vdev_start() 1193 cpu_to_le32(arg->ml.partner_info[i].vdev_id); in ath12k_wmi_vdev_start() 1195 cpu_to_le32(arg->ml.partner_info[i].hw_link_id); in ath12k_wmi_vdev_start() [all …]
|
| /linux/Documentation/bpf/ |
| H A D | map_queue_stack.rst | 146 https://lwn.net/ml/netdev/153986858555.9127.14517764371945179514.stgit@kernel/
|
| H A D | map_sk_storage.rst | 159 https://lwn.net/ml/netdev/20190426171103.61892-1-kafai@fb.com/
|
| H A D | map_bloom_filter.rst | 174 https://lwn.net/ml/bpf/20210831225005.2762202-1-joannekoong@fb.com/
|
| /linux/tools/perf/pmu-events/ |
| H A D | intel_metrics.py | 699 ml = d_ratio(l1d, l1dc) 702 ml, "load_miss_pending/cycle")
|