| /linux/drivers/spi/ |
| H A D | spi-mpc52xx.c | 87 int (*state)(int irq, struct mpc52xx_spi *ms, u8 status, u8 data); 100 static void mpc52xx_spi_chipsel(struct mpc52xx_spi *ms, int value) in mpc52xx_spi_chipsel() argument 104 if (ms->gpio_cs_count > 0) { in mpc52xx_spi_chipsel() 105 cs = spi_get_chipselect(ms->message->spi, 0); in mpc52xx_spi_chipsel() 106 gpiod_set_value(ms->gpio_cs[cs], value); in mpc52xx_spi_chipsel() 108 out_8(ms->regs + SPI_PORTDATA, value ? 0 : 0x08); in mpc52xx_spi_chipsel() 117 static void mpc52xx_spi_start_transfer(struct mpc52xx_spi *ms) in mpc52xx_spi_start_transfer() argument 119 ms->rx_buf = ms->transfer->rx_buf; in mpc52xx_spi_start_transfer() 120 ms->tx_buf = ms->transfer->tx_buf; in mpc52xx_spi_start_transfer() 121 ms->len = ms->transfer->len; in mpc52xx_spi_start_transfer() [all …]
|
| H A D | spi-mtk-snfi.c | 645 struct mtk_snand *ms = nand_to_mtk_snand(nand); in mtk_snand_ooblayout_free() local 647 if (section >= ms->nfi_cfg.nsectors) in mtk_snand_ooblayout_free() 650 oobfree->length = ms->caps->fdm_size - 1; in mtk_snand_ooblayout_free() 651 oobfree->offset = section * ms->caps->fdm_size + 1; in mtk_snand_ooblayout_free() 1257 struct mtk_snand *ms = spi_controller_get_devdata(mem->spi->controller); in mtk_snand_adjust_op_size() local 1265 if (ms->autofmt) in mtk_snand_adjust_op_size() 1267 l = ms->caps->sector_size + ms->nfi_cfg.spare_size; in mtk_snand_adjust_op_size() 1268 l *= ms->nfi_cfg.nsectors; in mtk_snand_adjust_op_size() 1284 struct mtk_snand *ms = spi_controller_get_devdata(mem->spi->controller); in mtk_snand_exec_op() local 1288 return mtk_snand_read_page_cache(ms, op); in mtk_snand_exec_op() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-raid1.c | 51 struct mirror_set *ms; member 99 struct mirror_set *ms = context; in wakeup_mirrord() local 101 queue_work(ms->kmirrord_wq, &ms->kmirrord_work); in wakeup_mirrord() 106 struct mirror_set *ms = timer_container_of(ms, t, timer); in delayed_wake_fn() local 108 clear_bit(0, &ms->timer_pending); in delayed_wake_fn() 109 wakeup_mirrord(ms); in delayed_wake_fn() 112 static void delayed_wake(struct mirror_set *ms) in delayed_wake() argument 114 if (test_and_set_bit(0, &ms->timer_pending)) in delayed_wake() 117 ms->timer.expires = jiffies + HZ / 5; in delayed_wake() 118 add_timer(&ms->timer); in delayed_wake() [all …]
|
| /linux/drivers/accel/ivpu/ |
| H A D | ivpu_ms.c | 24 struct ivpu_ms_instance *ms; in get_instance_by_mask() local 28 list_for_each_entry(ms, &file_priv->ms_instance_list, ms_instance_node) in get_instance_by_mask() 29 if (ms->mask == metric_mask) in get_instance_by_mask() 30 return ms; in get_instance_by_mask() 40 struct ivpu_ms_instance *ms; in ivpu_ms_start_ioctl() local 62 ms = kzalloc_obj(*ms); in ivpu_ms_start_ioctl() 63 if (!ms) { in ivpu_ms_start_ioctl() 68 ms->mask = args->metric_group_mask; in ivpu_ms_start_ioctl() 70 ret = ivpu_jsm_metric_streamer_info(vdev, ms->mask, 0, 0, &sample_size, NULL); in ivpu_ms_start_ioctl() 83 ms->bo = ivpu_bo_create_global(vdev, buf_size, DRM_IVPU_BO_CACHED | DRM_IVPU_BO_MAPPABLE); in ivpu_ms_start_ioctl() [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/dml2_0/ |
| H A D | display_mode_core.c | 6223 CalculatePrefetchSchedule_params->DSCDelay = mode_lib->ms.DSCDelayPerState[k]; in set_calculate_prefetch_schedule_params() 6224 …CalculatePrefetchSchedule_params->EnhancedPrefetchScheduleAccelerationFinal = mode_lib->ms.policy.… in set_calculate_prefetch_schedule_params() 6225 …arams->DPPCLKDelaySubtotalPlusCNVCFormater = mode_lib->ms.ip.dppclk_delay_subtotal + mode_lib->ms.… in set_calculate_prefetch_schedule_params() 6226 CalculatePrefetchSchedule_params->DPPCLKDelaySCL = mode_lib->ms.ip.dppclk_delay_scl; in set_calculate_prefetch_schedule_params() 6227 … CalculatePrefetchSchedule_params->DPPCLKDelaySCLLBOnly = mode_lib->ms.ip.dppclk_delay_scl_lb_only; in set_calculate_prefetch_schedule_params() 6228 …CalculatePrefetchSchedule_params->DPPCLKDelayCNVCCursor = mode_lib->ms.ip.dppclk_delay_cnvc_cursor; in set_calculate_prefetch_schedule_params() 6229 CalculatePrefetchSchedule_params->DISPCLKDelaySubtotal = mode_lib->ms.ip.dispclk_delay_subtotal; in set_calculate_prefetch_schedule_params() 6230 …edule_params->DPP_RECOUT_WIDTH = (dml_uint_t)(mode_lib->ms.SwathWidthYThisState[k] / mode_lib->ms.… in set_calculate_prefetch_schedule_params() 6231 …CalculatePrefetchSchedule_params->OutputFormat = mode_lib->ms.cache_display_cfg.output.OutputForma… in set_calculate_prefetch_schedule_params() 6232 …CalculatePrefetchSchedule_params->MaxInterDCNTileRepeaters = mode_lib->ms.ip.max_inter_dcn_tile_re… in set_calculate_prefetch_schedule_params() [all …]
|
| H A D | display_mode_util.c | 358 …dml_print("DML: MODE SUPPORT: Mode Supported : %s\n", mode_lib->ms.support.ModeSu… in dml_print_mode_support() 359 …dml_print("DML: MODE SUPPORT: Scale Ratio And Taps : %s\n", mode_lib->ms.suppor… in dml_print_mode_support() 360 …dml_print("DML: MODE SUPPORT: Source Format Pixel And Scan : %s\n", mode_lib->ms.suppor… in dml_print_mode_support() 361 …dml_print("DML: MODE SUPPORT: Viewport Size : %s\n", mode_lib->ms.suppor… in dml_print_mode_support() 362 …dml_print("DML: MODE SUPPORT: Link Rate Does Not Match DP Version : %s\n", mode_lib->ms… in dml_print_mode_support() 363 …dml_print("DML: MODE SUPPORT: Link Rate For Multistream Not Indicated : %s\n", mode_lib->ms… in dml_print_mode_support() 364 …dml_print("DML: MODE SUPPORT: BPP For Multi stream Not Indicated : %s\n", mode_lib->ms… in dml_print_mode_support() 365 …dml_print("DML: MODE SUPPORT: Multistream With HDMI Or eDP : %s\n", mode_lib->ms… in dml_print_mode_support() 366 …dml_print("DML: MODE SUPPORT: Exceeded Multistream Slots : %s\n", mode_lib->ms… in dml_print_mode_support() 367 …dml_print("DML: MODE SUPPORT: MSO Or ODM Split With Non DP Link : %s\n", mode_lib->ms… in dml_print_mode_support() [all …]
|
| /linux/Documentation/fb/ |
| H A D | viafb.modes | 15 # Sync Width 3.813 us 0.064 ms 17 # Front Porch 0.636 us 0.318 ms 19 # Back Porch 1.907 us 1.048 ms 21 # Active Time 25.422 us 15.253 ms 23 # Blank Time 6.356 us 1.430 ms 40 # Sync Width 2.032 us 0.080 ms 42 # Front Porch 0.508 us 0.027 ms 44 # Back Porch 3.810 us 0.427 ms 46 # Active Time 20.317 us 12.800 ms 48 # Blank Time 6.349 us 0.533 ms [all …]
|
| /linux/drivers/usb/typec/ |
| H A D | mode_selection.c | 104 struct mode_state *ms, *tmp; in mode_list_clean() local 106 list_for_each_entry_safe(ms, tmp, &sel->mode_list, list) { in mode_list_clean() 107 list_del(&ms->list); in mode_list_clean() 108 kfree(ms); in mode_list_clean() 134 struct mode_state *ms; in mode_selection_work_fn() local 140 ms = list_first_entry_or_null(&sel->mode_list, struct mode_state, list); in mode_selection_work_fn() 141 if (!ms) in mode_selection_work_fn() 144 if (sel->active_svid == ms->svid) { in mode_selection_work_fn() 145 dev_dbg(&sel->partner->dev, "%x altmode is active\n", ms->svid); in mode_selection_work_fn() 153 } else if (ms->error) { in mode_selection_work_fn() [all …]
|
| /linux/Documentation/translations/zh_CN/scheduler/ |
| H A D | sched-bwc.rst | 78 cpu.cfs_period_us=100ms 86 小配额是1ms。周期长度也有一个1s的上限。当带宽限制以分层方式使用时,存在额外的限制,这些在下 104 /proc/sys/kernel/sched_cfs_bandwidth_slice_us (default=5ms) 144 1ms以外的所有时间片都可以返回到全局池中。这是在编译时由min_cfs_rq_runtime变量配置的。这 154 即任务组正在运行的每个cpu上未使用的片断量(通常每个cpu最多1ms或由min_cfs_rq_runtime定 157 长的时间窗口。这也限制了突发能力,每个cpu不超过1ms。这为在高核数机器上有小配额限制的高线程 164 可能在某些时期使用多达1ms的额外配额,从而阻止cpu绑定的应用程序完全使用其配额,这也是同样的数 172 如果周期是250ms,配额也是250ms,那么该组将每250ms获得价值1个CPU的运行时间。 174 # echo 250000 > cpu.cfs_quota_us /* quota = 250ms */ 175 # echo 250000 > cpu.cfs_period_us /* period = 250ms */ [all …]
|
| /linux/lib/zstd/compress/ |
| H A D | zstd_lazy.c | 30 void ZSTD_updateDUBT(ZSTD_MatchState_t* ms, in ZSTD_updateDUBT() argument 34 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_updateDUBT() 35 U32* const hashTable = ms->hashTable; in ZSTD_updateDUBT() 38 U32* const bt = ms->chainTable; in ZSTD_updateDUBT() 42 const BYTE* const base = ms->window.base; in ZSTD_updateDUBT() 44 U32 idx = ms->nextToUpdate; in ZSTD_updateDUBT() 48 idx, target, ms->window.dictLimit); in ZSTD_updateDUBT() 52 assert(idx >= ms->window.dictLimit); /* condition for valid base+idx */ in ZSTD_updateDUBT() 65 ms->nextToUpdate = target; in ZSTD_updateDUBT() 75 void ZSTD_insertDUBT1(const ZSTD_MatchState_t* ms, in ZSTD_insertDUBT1() argument [all …]
|
| H A D | zstd_lazy.h | 31 U32 ZSTD_insertAndFindFirstIndex(ZSTD_MatchState_t* ms, const BYTE* ip); 32 void ZSTD_row_update(ZSTD_MatchState_t* const ms, const BYTE* ip); 34 void ZSTD_dedicatedDictSearch_lazy_loadDictionary(ZSTD_MatchState_t* ms, const BYTE* const ip); 41 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 44 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 47 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 50 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 53 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 56 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], 59 ZSTD_MatchState_t* ms, SeqStore_t* seqStore, U32 rep[ZSTD_REP_NUM], [all …]
|
| H A D | zstd_fast.c | 17 void ZSTD_fillHashTableForCDict(ZSTD_MatchState_t* ms, in ZSTD_fillHashTableForCDict() argument 21 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_fillHashTableForCDict() 22 U32* const hashTable = ms->hashTable; in ZSTD_fillHashTableForCDict() 25 const BYTE* const base = ms->window.base; in ZSTD_fillHashTableForCDict() 26 const BYTE* ip = base + ms->nextToUpdate; in ZSTD_fillHashTableForCDict() 54 void ZSTD_fillHashTableForCCtx(ZSTD_MatchState_t* ms, in ZSTD_fillHashTableForCCtx() argument 58 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_fillHashTableForCCtx() 59 U32* const hashTable = ms->hashTable; in ZSTD_fillHashTableForCCtx() 62 const BYTE* const base = ms->window.base; in ZSTD_fillHashTableForCCtx() 63 const BYTE* ip = base + ms->nextToUpdate; in ZSTD_fillHashTableForCCtx() [all …]
|
| H A D | zstd_double_fast.c | 19 void ZSTD_fillDoubleHashTableForCDict(ZSTD_MatchState_t* ms, in ZSTD_fillDoubleHashTableForCDict() argument 22 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_fillDoubleHashTableForCDict() 23 U32* const hashLarge = ms->hashTable; in ZSTD_fillDoubleHashTableForCDict() 26 U32* const hashSmall = ms->chainTable; in ZSTD_fillDoubleHashTableForCDict() 28 const BYTE* const base = ms->window.base; in ZSTD_fillDoubleHashTableForCDict() 29 const BYTE* ip = base + ms->nextToUpdate; in ZSTD_fillDoubleHashTableForCDict() 57 void ZSTD_fillDoubleHashTableForCCtx(ZSTD_MatchState_t* ms, in ZSTD_fillDoubleHashTableForCCtx() argument 60 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_fillDoubleHashTableForCCtx() 61 U32* const hashLarge = ms->hashTable; in ZSTD_fillDoubleHashTableForCCtx() 64 U32* const hashSmall = ms->chainTable; in ZSTD_fillDoubleHashTableForCCtx() [all …]
|
| H A D | zstd_opt.c | 412 U32 ZSTD_insertAndFindFirstIndexHash3 (const ZSTD_MatchState_t* ms, in ZSTD_insertAndFindFirstIndexHash3() argument 416 U32* const hashTable3 = ms->hashTable3; in ZSTD_insertAndFindFirstIndexHash3() 417 U32 const hashLog3 = ms->hashLog3; in ZSTD_insertAndFindFirstIndexHash3() 418 const BYTE* const base = ms->window.base; in ZSTD_insertAndFindFirstIndexHash3() 444 const ZSTD_MatchState_t* ms, in ZSTD_insertBt1() argument 449 const ZSTD_compressionParameters* const cParams = &ms->cParams; in ZSTD_insertBt1() 450 U32* const hashTable = ms->hashTable; in ZSTD_insertBt1() 453 U32* const bt = ms->chainTable; in ZSTD_insertBt1() 458 const BYTE* const base = ms->window.base; in ZSTD_insertBt1() 459 const BYTE* const dictBase = ms->window.dictBase; in ZSTD_insertBt1() [all …]
|
| /linux/fs/dlm/ |
| H A D | requestqueue.c | 34 const struct dlm_message *ms) in dlm_add_requestqueue() argument 37 int length = le16_to_cpu(ms->m_header.h_length) - in dlm_add_requestqueue() 48 memcpy(&e->request, ms, sizeof(*ms)); in dlm_add_requestqueue() 49 memcpy(&e->request.m_extra, ms->m_extra, length); in dlm_add_requestqueue() 68 struct dlm_message *ms; in dlm_process_requestqueue() local 80 ms = &e->request; in dlm_process_requestqueue() 84 le32_to_cpu(ms->m_type), in dlm_process_requestqueue() 85 le32_to_cpu(ms->m_header.h_nodeid), in dlm_process_requestqueue() 86 le32_to_cpu(ms->m_lkid), le32_to_cpu(ms->m_remid), in dlm_process_requestqueue() 87 from_dlm_errno(le32_to_cpu(ms->m_result)), in dlm_process_requestqueue() [all …]
|
| H A D | lock.c | 89 const struct dlm_message *ms, bool local); 90 static int receive_extralen(const struct dlm_message *ms); 1772 const struct dlm_message *ms) in _remove_from_waiters() argument 1810 (lkb->lkb_wait_type == DLM_MSG_CONVERT) && ms && !ms->m_result && in _remove_from_waiters() 1829 lkb->lkb_id, ms ? le32_to_cpu(ms->m_header.h_nodeid) : 0, in _remove_from_waiters() 1876 const struct dlm_message *ms, bool local) in remove_from_waiters_ms() argument 1886 error = _remove_from_waiters(lkb, le32_to_cpu(ms->m_type), ms); in remove_from_waiters_ms() 1975 const struct dlm_message *ms) in set_lvb_lock_pc() argument 1987 int len = receive_extralen(ms); in set_lvb_lock_pc() 1990 memcpy(lkb->lkb_lvbptr, ms->m_extra, len); in set_lvb_lock_pc() [all …]
|
| /linux/kernel/ |
| H A D | module_signature.c | 21 int mod_check_sig(const struct module_signature *ms, size_t file_len, in mod_check_sig() argument 24 if (be32_to_cpu(ms->sig_len) >= file_len - sizeof(*ms)) in mod_check_sig() 27 if (ms->id_type != PKEY_ID_PKCS7) { in mod_check_sig() 33 if (ms->algo != 0 || in mod_check_sig() 34 ms->hash != 0 || in mod_check_sig() 35 ms->signer_len != 0 || in mod_check_sig() 36 ms->key_id_len != 0 || in mod_check_sig() 37 ms->__pad[0] != 0 || in mod_check_sig() 38 ms->__pad[1] != 0 || in mod_check_sig() 39 ms->__pad[2] != 0) { in mod_check_sig()
|
| /linux/include/trace/events/ |
| H A D | dlm.h | 393 TP_PROTO(uint32_t dst, uint32_t h_seq, const struct dlm_message *ms, 396 TP_ARGS(dst, h_seq, ms, name, namelen), 425 le16_to_cpu(ms->m_header.h_length) - sizeof(*ms)) 432 __entry->h_version = le32_to_cpu(ms->m_header.h_version); 433 __entry->h_lockspace = le32_to_cpu(ms->m_header.u.h_lockspace); 434 __entry->h_nodeid = le32_to_cpu(ms->m_header.h_nodeid); 435 __entry->h_length = le16_to_cpu(ms->m_header.h_length); 436 __entry->h_cmd = ms->m_header.h_cmd; 437 __entry->m_type = le32_to_cpu(ms->m_type); 438 __entry->m_nodeid = le32_to_cpu(ms->m_nodeid); [all …]
|
| /linux/drivers/hid/ |
| H A D | hid-microsoft.c | 62 struct ms_data *ms = hid_get_drvdata(hdev); in ms_report_fixup() local 63 unsigned long quirks = ms->quirks; in ms_report_fixup() 186 struct ms_data *ms = hid_get_drvdata(hdev); in ms_input_mapping() local 187 unsigned long quirks = ms->quirks; in ms_input_mapping() 213 struct ms_data *ms = hid_get_drvdata(hdev); in ms_input_mapped() local 214 unsigned long quirks = ms->quirks; in ms_input_mapped() 225 struct ms_data *ms = hid_get_drvdata(hdev); in ms_event() local 226 unsigned long quirks = ms->quirks; in ms_event() 283 struct ms_data *ms = container_of(work, struct ms_data, ff_worker); in ms_ff_worker() local 284 struct hid_device *hdev = ms->hdev; in ms_ff_worker() [all …]
|
| /linux/include/asm-generic/ |
| H A D | mmiowb.h | 37 struct mmiowb_state *ms = __mmiowb_state(); in mmiowb_set_pending() local 39 if (likely(ms->nesting_count)) in mmiowb_set_pending() 40 ms->mmiowb_pending = ms->nesting_count; in mmiowb_set_pending() 45 struct mmiowb_state *ms = __mmiowb_state(); in mmiowb_spin_lock() local 46 ms->nesting_count++; in mmiowb_spin_lock() 51 struct mmiowb_state *ms = __mmiowb_state(); in mmiowb_spin_unlock() local 53 if (unlikely(ms->mmiowb_pending)) { in mmiowb_spin_unlock() 54 ms->mmiowb_pending = 0; in mmiowb_spin_unlock() 58 ms->nesting_count--; in mmiowb_spin_unlock()
|
| /linux/drivers/platform/x86/amd/pmf/ |
| H A D | cnqf.c | 273 struct cnqf_mode_settings *ms; in amd_pmf_update_mode_set() local 276 ms = &config_store.mode_set[idx][CNQF_MODE_QUIET]; in amd_pmf_update_mode_set() 277 ms->power_floor = out->ps[APMF_CNQF_QUIET].pfloor; in amd_pmf_update_mode_set() 278 ms->power_control.fppt = out->ps[APMF_CNQF_QUIET].fppt; in amd_pmf_update_mode_set() 279 ms->power_control.sppt = out->ps[APMF_CNQF_QUIET].sppt; in amd_pmf_update_mode_set() 280 ms->power_control.sppt_apu_only = out->ps[APMF_CNQF_QUIET].sppt_apu_only; in amd_pmf_update_mode_set() 281 ms->power_control.spl = out->ps[APMF_CNQF_QUIET].spl; in amd_pmf_update_mode_set() 282 ms->power_control.stt_min = out->ps[APMF_CNQF_QUIET].stt_min_limit; in amd_pmf_update_mode_set() 283 ms->power_control.stt_skin_temp[STT_TEMP_APU] = in amd_pmf_update_mode_set() 285 ms->power_control.stt_skin_temp[STT_TEMP_HS2] = in amd_pmf_update_mode_set() [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/ |
| H A D | dml2_core_dcn4_calcs.c | 283 dml_get_per_plane_var_func(num_mcaches_plane0, unsigned int, mode_lib->ms.num_mcaches_l); 284 dml_get_per_plane_var_func(mcache_row_bytes_plane0, unsigned int, mode_lib->ms.mcache_row_bytes_l); 285 dml_get_per_plane_var_func(mcache_shift_granularity_plane0, unsigned int, mode_lib->ms.mcache_shift… 286 dml_get_per_plane_var_func(num_mcaches_plane1, unsigned int, mode_lib->ms.num_mcaches_c); 287 dml_get_per_plane_var_func(mcache_row_bytes_plane1, unsigned int, mode_lib->ms.mcache_row_bytes_c); 288 dml_get_per_plane_var_func(mcache_shift_granularity_plane1, unsigned int, mode_lib->ms.mcache_shift… 289 dml_get_per_plane_var_func(mall_comb_mcache_l, unsigned int, mode_lib->ms.mall_comb_mcache_l); 290 dml_get_per_plane_var_func(mall_comb_mcache_c, unsigned int, mode_lib->ms.mall_comb_mcache_c); 291 dml_get_per_plane_var_func(lc_comb_mcache, unsigned int, mode_lib->ms.lc_comb_mcache); 292 dml_get_per_plane_var_func(subviewport_lines_needed_in_mall, unsigned int, mode_lib->ms.SubViewport… [all …]
|
| /linux/tools/perf/util/annotate-arch/ |
| H A D | annotate-loongarch.c | 18 struct map_symbol *ms, in loongarch_call__parse() argument 22 struct map *map = ms->map; in loongarch_call__parse() 50 .ms = { .map = map__get(map), }, in loongarch_call__parse() 54 if (maps__find_ams(thread__maps(ms->thread), &target) == 0 && in loongarch_call__parse() 55 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr) in loongarch_call__parse() 56 ops->target.sym = target.ms.sym; in loongarch_call__parse() 69 struct map_symbol *ms, in loongarch_jump__parse() argument 73 struct map *map = ms->map; in loongarch_jump__parse() 74 struct symbol *sym = ms->sym; in loongarch_jump__parse() 76 .ms = { .map = map__get(map), }, in loongarch_jump__parse() [all …]
|
| /linux/net/mac80211/ |
| H A D | rc80211_minstrel_ht_debugfs.c | 22 struct minstrel_debugfs_info *ms; in minstrel_stats_read() local 24 ms = file->private_data; in minstrel_stats_read() 25 return simple_read_from_buffer(buf, len, ppos, ms->buf, ms->len); in minstrel_stats_read() 148 struct minstrel_debugfs_info *ms; in minstrel_ht_stats_open() local 152 ms = kmalloc(32768, GFP_KERNEL); in minstrel_ht_stats_open() 153 if (!ms) in minstrel_ht_stats_open() 156 file->private_data = ms; in minstrel_ht_stats_open() 157 p = ms->buf; in minstrel_ht_stats_open() 179 ms->len = p - ms->buf; in minstrel_ht_stats_open() 180 WARN_ON(ms->len + sizeof(*ms) > 32768); in minstrel_ht_stats_open() [all …]
|
| /linux/drivers/scsi/bfa/ |
| H A D | bfa_fcs_lport.c | 1435 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_offline() 1475 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_sending_rhba() 1500 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rhba() 1549 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rhba_retry() 1580 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_sending_rprt() 1605 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rprt() 1652 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rprt_retry() 1683 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_sending_rpa() 1708 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rpa() 1753 struct bfa_fcs_lport_s *port = fdmi->ms->port; in bfa_fcs_lport_fdmi_sm_rpa_retry() [all …]
|