Searched refs:devtype_data (Results 1 – 13 of 13) sorted by relevance
186 const struct mxc_nand_devtype_data *devtype_data; member258 u16 sparebuf_size = host->devtype_data->spare_len; in copy_spare()308 if (!host->devtype_data->irqpending_quirk) in check_int_v1_v2()344 if (host->devtype_data->irqpending_quirk) { in irq_control()350 host->devtype_data->irq_control(host, activate); in irq_control()435 if (!host->devtype_data->check_int(host)) in mxc_nfc_irq()456 if (host->devtype_data->check_int(host)) in wait_op_done()467 if (!time_left && !host->devtype_data->check_int(host)) { in wait_op_done()478 done = host->devtype_data->check_int(host); in wait_op_done()516 if (host->devtype_data->irqpending_quirk && (cmd == NAND_CMD_RESET)) { in send_cmd_v1_v2()[all …]
243 static const struct fsl_dspi_devtype_data devtype_data[] = { variable356 const struct fsl_dspi_devtype_data *devtype_data; member382 return data->devtype_data == &devtype_data[S32G] || in is_s32g_dspi()383 data->devtype_data == &devtype_data[S32G_TARGET]; in is_s32g_dspi()962 if (dspi->devtype_data->trans_mode == DSPI_DMA_MODE || in dspi_setup_accel()966 if (!odd && dspi->len <= dspi->devtype_data->fifo_size * 2) { in dspi_setup_accel()968 } else if (odd && dspi->len <= dspi->devtype_data->fifo_size) { in dspi_setup_accel()1016 int num_fifo_entries = dspi->devtype_data->fifo_size; in dspi_fifo_write()1217 if (dspi->devtype_data->trans_mode == DSPI_DMA_MODE) { in dspi_transfer_one_message()1395 .data = &devtype_data[VF610],[all …]
148 const struct spi_imx_devtype_data *devtype_data; member153 return d->devtype_data->devtype == IMX27_CSPI; in is_imx27_cspi()158 return d->devtype_data->devtype == IMX35_CSPI; in is_imx35_cspi()163 return d->devtype_data->devtype == IMX51_ECSPI; in is_imx51_ecspi()168 return d->devtype_data->devtype == IMX53_ECSPI; in is_imx53_ecspi()276 if (transfer->len < spi_imx->devtype_data->fifo_size) in spi_imx_can_dma()754 if (spi_imx->usedma && spi_imx->devtype_data->tx_glitch_fixed) in mx51_ecspi_prepare_transfer()811 if (spi_imx->devtype_data->tx_glitch_fixed) in mx51_setup_wml()1272 while (spi_imx->txfifo < spi_imx->devtype_data->fifo_size) { in spi_imx_push()1283 spi_imx->devtype_data->trigger(spi_imx); in spi_imx_push()[all …]
108 struct flexcan_devtype_data devtype_data; member130 const u32 quirks = priv->devtype_data.quirks; in flexcan_supports_rx_mailbox()138 const u32 quirks = priv->devtype_data.quirks; in flexcan_supports_rx_mailbox_rtr()149 const u32 quirks = priv->devtype_data.quirks; in flexcan_supports_rx_fifo()157 const u32 quirks = priv->devtype_data.quirks; in flexcan_active_rx_rtr()
30 if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) in flexcan_get_ringparam()64 u32 quirks = priv->devtype_data.quirks; in flexcan_set_priv_flags()80 if (quirks != priv->devtype_data.quirks && netif_running(ndev)) in flexcan_set_priv_flags()83 priv->devtype_data.quirks = quirks; in flexcan_set_priv_flags()
92 const struct imx_ddr_devtype_data *devtype_data; member119 return pmu->devtype_data->filter_ver == DDR_PERF_AXI_FILTER_V1; in axi_filter_v1()124 return pmu->devtype_data->filter_ver == DDR_PERF_AXI_FILTER_V2; in axi_filter_v2()142 return sysfs_emit(page, "%s\n", pmu->devtype_data->identifier); in ddr_perf_identifier_show()180 const struct imx_ddr_devtype_data *devtype_data; member199 .devtype_data = _data, } \329 if (!eattr->devtype_data) in ddr_perf_events_attrs_is_visible()332 if (eattr->devtype_data != ddr_pmu->devtype_data && in ddr_perf_events_attrs_is_visible()333 eattr->devtype_data->filter_ver != ddr_pmu->devtype_data->filter_ver) in ddr_perf_events_attrs_is_visible()796 pmu->devtype_data = of_device_get_match_data(&pdev->dev); in ddr_perf_probe()
160 struct at91_devtype_data devtype_data; member202 return priv->devtype_data.type == AT91_DEVTYPE_SAM##_model; \210 return priv->devtype_data.rx_first; in get_mb_rx_first()215 return priv->devtype_data.rx_last; in get_mb_rx_last()220 return priv->devtype_data.tx_shift; in get_mb_tx_shift()1048 const struct at91_devtype_data *devtype_data; in at91_can_probe() local1057 devtype_data = at91_can_get_driver_data(pdev); in at91_can_probe()1058 if (!devtype_data) { in at91_can_probe()1092 1 << devtype_data->tx_shift); in at91_can_probe()1118 priv->devtype_data = *devtype_data; in at91_can_probe()[all …]
38 if (priv->devtype_data.quirks & RKCANFD_QUIRK_RK3568_ERRATUM_6 && in rkcanfd_get_effective_tx_free()48 if (priv->devtype_data.quirks & RKCANFD_QUIRK_RK3568_ERRATUM_12) in rkcanfd_start_xmit_write_cmd()54 if (priv->devtype_data.quirks & RKCANFD_QUIRK_RK3568_ERRATUM_12) in rkcanfd_start_xmit_write_cmd()
135 if (!(priv->devtype_data.quirks & RKCANFD_QUIRK_RK3568_ERRATUM_6)) in rkcanfd_rxstx_filter()
471 struct rkcanfd_devtype_data devtype_data; member
115 return __mcp251xfd_get_model_str(priv->devtype_data.model); in mcp251xfd_get_model_str()649 if (priv->devtype_data.quirks & MCP251XFD_QUIRK_ECC) in mcp251xfd_chip_ecc_init()1159 if ((priv->devtype_data.quirks & MCP251XFD_QUIRK_MAB_NO_WARN) && in mcp251xfd_handle_modif()1226 if (priv->devtype_data.quirks & MCP251XFD_QUIRK_MAB_NO_WARN) in mcp251xfd_handle_serrif()1717 priv->devtype_data.quirks |= MCP251XFD_QUIRK_HALF_DUPLEX; in mcp251xfd_register_quirks()1723 const struct mcp251xfd_devtype_data *devtype_data; in mcp251xfd_register_chip_detect() local1746 devtype_data = &mcp251xfd_devtype_data_mcp251863; in mcp251xfd_register_chip_detect()1748 devtype_data = &mcp251xfd_devtype_data_mcp2518fd; in mcp251xfd_register_chip_detect()1750 devtype_data = &mcp251xfd_devtype_data_mcp2517fd; in mcp251xfd_register_chip_detect()1754 priv->devtype_data.model != devtype_data->model) { in mcp251xfd_register_chip_detect()[all …]
88 if (priv->devtype_data.quirks & MCP251XFD_QUIRK_CRC_TX) in mcp251xfd_tx_obj_from_skb()115 if (priv->devtype_data.quirks & MCP251XFD_QUIRK_CRC_TX) { in mcp251xfd_tx_obj_from_skb()
107 const struct fsl_mmdc_devtype_data *devtype_data; member355 if (pmu_mmdc->devtype_data->flags & MMDC_FLAG_PROFILE_SEL) in mmdc_pmu_event_start()510 pmu_mmdc->devtype_data = device_get_match_data(&pdev->dev); in imx_mmdc_perf_init()