/linux/drivers/crypto/ |
H A D | omap-des.c | 43 #define _calc_walked(inout) (dd->inout##_walk.offset - dd->inout##_sg->offset) 45 #define DES_REG_KEY(dd, x) ((dd)->pdata->key_ofs - \ argument 48 #define DES_REG_IV(dd, x) ((dd)->pdata->iv_ofs + ((x) * 0x04)) argument 50 #define DES_REG_CTRL(dd) ((dd)->pdata->ctrl_ofs) argument 57 #define DES_REG_DATA_N(dd, x) ((dd)->pdata->data_ofs + ((x) * 0x04)) argument 59 #define DES_REG_REV(dd) ((dd)->pdata->rev_ofs) argument 61 #define DES_REG_MASK(dd) ((dd)->pdata->mask_ofs) argument 65 #define DES_REG_IRQ_STATUS(dd) ((dd)->pdata->irq_status_ofs) argument 66 #define DES_REG_IRQ_ENABLE(dd) ((dd)->pdata->irq_enable_ofs) argument 82 struct omap_des_dev *dd; member [all …]
|
H A D | omap-aes.c | 47 #define omap_aes_read(dd, offset) \ argument 50 _read_ret = __raw_readl(dd->io_base + offset); \ 56 inline u32 omap_aes_read(struct omap_aes_dev *dd, u32 offset) in omap_aes_read() argument 58 return __raw_readl(dd->io_base + offset); in omap_aes_read() 63 #define omap_aes_write(dd, offset, value) \ argument 67 __raw_writel(value, dd->io_base + offset); \ 70 inline void omap_aes_write(struct omap_aes_dev *dd, u32 offset, in omap_aes_write() argument 73 __raw_writel(value, dd->io_base + offset); in omap_aes_write() 77 static inline void omap_aes_write_mask(struct omap_aes_dev *dd, u32 offset, in omap_aes_write_mask() argument 82 val = omap_aes_read(dd, offset); in omap_aes_write_mask() [all …]
|
H A D | atmel-tdes.c | 65 struct atmel_tdes_dev *dd; member 171 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset) in atmel_tdes_read() argument 173 return readl_relaxed(dd->io_base + offset); in atmel_tdes_read() 176 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd, in atmel_tdes_write() argument 179 writel_relaxed(value, dd->io_base + offset); in atmel_tdes_write() 182 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset, in atmel_tdes_write_n() argument 186 atmel_tdes_write(dd, offset, *value); in atmel_tdes_write_n() 201 static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd) in atmel_tdes_hw_init() argument 205 err = clk_prepare_enable(dd->iclk); in atmel_tdes_hw_init() 209 if (!(dd->flags & TDES_FLAGS_INIT)) { in atmel_tdes_hw_init() [all …]
|
H A D | atmel-aes.c | 92 struct atmel_aes_dev *dd; member 335 static inline u32 atmel_aes_read(struct atmel_aes_dev *dd, u32 offset) in atmel_aes_read() argument 337 u32 value = readl_relaxed(dd->io_base + offset); in atmel_aes_read() 340 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_read() 343 dev_vdbg(dd->dev, "read 0x%08x from %s\n", value, in atmel_aes_read() 351 static inline void atmel_aes_write(struct atmel_aes_dev *dd, in atmel_aes_write() argument 355 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_write() 358 dev_vdbg(dd->dev, "write 0x%08x into %s\n", value, in atmel_aes_write() 363 writel_relaxed(value, dd->io_base + offset); in atmel_aes_write() 366 static void atmel_aes_read_n(struct atmel_aes_dev *dd, u32 offset, in atmel_aes_read_n() argument [all …]
|
H A D | omap-aes-gcm.c | 29 static int omap_aes_gcm_handle_queue(struct omap_aes_dev *dd, 32 static void omap_aes_gcm_finish_req(struct omap_aes_dev *dd, int ret) in omap_aes_gcm_finish_req() argument 34 struct aead_request *req = dd->aead_req; in omap_aes_gcm_finish_req() 36 dd->in_sg = NULL; in omap_aes_gcm_finish_req() 37 dd->out_sg = NULL; in omap_aes_gcm_finish_req() 39 crypto_finalize_aead_request(dd->engine, req, ret); in omap_aes_gcm_finish_req() 41 pm_runtime_mark_last_busy(dd->dev); in omap_aes_gcm_finish_req() 42 pm_runtime_put_autosuspend(dd->dev); in omap_aes_gcm_finish_req() 45 static void omap_aes_gcm_done_task(struct omap_aes_dev *dd) in omap_aes_gcm_done_task() argument 51 alen = ALIGN(dd->assoc_len, AES_BLOCK_SIZE); in omap_aes_gcm_done_task() [all …]
|
H A D | atmel-sha.c | 90 struct atmel_sha_dev *dd; member 114 struct atmel_sha_dev *dd; member 253 static inline u32 atmel_sha_read(struct atmel_sha_dev *dd, u32 offset) in atmel_sha_read() argument 255 u32 value = readl_relaxed(dd->io_base + offset); in atmel_sha_read() 258 if (dd->flags & SHA_FLAGS_DUMP_REG) { in atmel_sha_read() 261 dev_vdbg(dd->dev, "read 0x%08x from %s\n", value, in atmel_sha_read() 269 static inline void atmel_sha_write(struct atmel_sha_dev *dd, in atmel_sha_write() argument 273 if (dd->flags & SHA_FLAGS_DUMP_REG) { in atmel_sha_write() 276 dev_vdbg(dd->dev, "write 0x%08x into %s\n", value, in atmel_sha_write() 281 writel_relaxed(value, dd->io_base + offset); in atmel_sha_write() [all …]
|
H A D | omap-sham.c | 43 #define SHA_REG_IDIGEST(dd, x) ((dd)->pdata->idigest_ofs + ((x)*0x04)) argument 44 #define SHA_REG_DIN(dd, x) ((dd)->pdata->din_ofs + ((x) * 0x04)) argument 45 #define SHA_REG_DIGCNT(dd) ((dd)->pdata->digcnt_ofs) argument 47 #define SHA_REG_ODIGEST(dd, x) ((dd)->pdata->odigest_ofs + (x * 0x04)) argument 57 #define SHA_REG_REV(dd) ((dd)->pdata->rev_ofs) argument 59 #define SHA_REG_MASK(dd) ((dd)->pdata->mask_ofs) argument 65 #define SHA_REG_SYSSTATUS(dd) ((dd)->pdata->sysstatus_ofs) argument 68 #define SHA_REG_MODE(dd) ((dd)->pdata->mode_ofs) argument 82 #define SHA_REG_LENGTH(dd) ((dd)->pdata->length_ofs) argument 141 struct omap_sham_dev *dd; member [all …]
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_iba6120.c | 306 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument 309 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32() 312 if (dd->userbase) in qib_read_ureg32() 314 ((char __iomem *)dd->userbase + in qib_read_ureg32() 315 dd->ureg_align * ctxt)); in qib_read_ureg32() 318 (dd->uregbase + in qib_read_ureg32() 319 (char __iomem *)dd->kregbase + in qib_read_ureg32() 320 dd->ureg_align * ctxt)); in qib_read_ureg32() 332 static inline void qib_write_ureg(const struct qib_devdata *dd, in qib_write_ureg() argument 337 if (dd->userbase) in qib_write_ureg() [all …]
|
H A D | qib_twsi.c | 67 static void i2c_wait_for_writes(struct qib_devdata *dd) in i2c_wait_for_writes() argument 74 dd->f_gpio_mod(dd, 0, 0, 0); in i2c_wait_for_writes() 89 static void scl_out(struct qib_devdata *dd, u8 bit) in scl_out() argument 95 mask = 1UL << dd->gpio_scl_num; in scl_out() 98 dd->f_gpio_mod(dd, 0, bit ? 0 : mask, mask); in scl_out() 110 if (mask & dd->f_gpio_mod(dd, 0, 0, 0)) in scl_out() 115 qib_dev_err(dd, "SCL interface stuck low > %d uSec\n", in scl_out() 118 i2c_wait_for_writes(dd); in scl_out() 121 static void sda_out(struct qib_devdata *dd, u8 bit) in sda_out() argument 125 mask = 1UL << dd->gpio_sda_num; in sda_out() [all …]
|
H A D | qib_tx.c | 61 void qib_disarm_piobufs(struct qib_devdata *dd, unsigned first, unsigned cnt) in qib_disarm_piobufs() argument 68 spin_lock_irqsave(&dd->pioavail_lock, flags); in qib_disarm_piobufs() 70 __clear_bit(i, dd->pio_need_disarm); in qib_disarm_piobufs() 71 dd->f_sendctrl(dd->pport, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs() 73 spin_unlock_irqrestore(&dd->pioavail_lock, flags); in qib_disarm_piobufs() 82 struct qib_devdata *dd = rcd->dd; in qib_disarm_piobufs_ifneeded() local 102 spin_lock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded() 104 if (__test_and_clear_bit(i, dd->pio_need_disarm)) in qib_disarm_piobufs_ifneeded() 105 dd->f_sendctrl(rcd->ppd, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs_ifneeded() 107 spin_unlock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded() [all …]
|
H A D | qib_sd7220.c | 98 static int qib_sd7220_reg_mod(struct qib_devdata *dd, int sdnum, u32 loc, 100 static int ibsd_mod_allchnls(struct qib_devdata *dd, int loc, int val, 102 static int qib_sd_trimdone_poll(struct qib_devdata *dd); 103 static void qib_sd_trimdone_monitor(struct qib_devdata *dd, const char *where); 104 static int qib_sd_setvals(struct qib_devdata *dd); 105 static int qib_sd_early(struct qib_devdata *dd); 106 static int qib_sd_dactrim(struct qib_devdata *dd); 107 static int qib_internal_presets(struct qib_devdata *dd); 109 static int qib_sd_trimself(struct qib_devdata *dd, int val); 110 static int epb_access(struct qib_devdata *dd, int sdnum, int claim); [all …]
|
H A D | qib_diag.c | 69 struct qib_devdata *dd; member 78 static struct qib_diag_client *get_client(struct qib_devdata *dd) in get_client() argument 92 dc->dd = dd; in get_client() 104 struct qib_devdata *dd = dc->dd; in return_client() local 108 if (dc == dd->diag_client) { in return_client() 109 dd->diag_client = dc->next; in return_client() 112 tdc = dc->dd->diag_client; in return_client() 124 rdc->dd = NULL; in return_client() 160 int qib_diag_add(struct qib_devdata *dd) in qib_diag_add() argument 173 snprintf(name, sizeof(name), "ipath_diag%d", dd->unit); in qib_diag_add() [all …]
|
H A D | qib_eeprom.c | 53 int qib_eeprom_read(struct qib_devdata *dd, u8 eeprom_offset, in qib_eeprom_read() argument 58 ret = mutex_lock_interruptible(&dd->eep_lock); in qib_eeprom_read() 60 ret = qib_twsi_reset(dd); in qib_eeprom_read() 62 qib_dev_err(dd, "EEPROM Reset for read failed\n"); in qib_eeprom_read() 64 ret = qib_twsi_blk_rd(dd, dd->twsi_eeprom_dev, in qib_eeprom_read() 66 mutex_unlock(&dd->eep_lock); in qib_eeprom_read() 77 static int eeprom_write_with_enable(struct qib_devdata *dd, u8 offset, in eeprom_write_with_enable() argument 82 pwen = dd->f_eeprom_wen(dd, 1); in eeprom_write_with_enable() 83 ret = qib_twsi_reset(dd); in eeprom_write_with_enable() 85 qib_dev_err(dd, "EEPROM Reset for write failed\n"); in eeprom_write_with_enable() [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | pcie.c | 24 int hfi1_pcie_init(struct hfi1_devdata *dd) in hfi1_pcie_init() argument 27 struct pci_dev *pdev = dd->pcidev; in hfi1_pcie_init() 43 dd_dev_err(dd, "pci enable failed: error %d\n", -ret); in hfi1_pcie_init() 49 dd_dev_err(dd, "pci_request_regions fails: err %d\n", -ret); in hfi1_pcie_init() 62 dd_dev_err(dd, "Unable to set DMA mask: %d\n", ret); in hfi1_pcie_init() 93 int hfi1_pcie_ddinit(struct hfi1_devdata *dd, struct pci_dev *pdev) in hfi1_pcie_ddinit() argument 110 dd_dev_err(dd, "chip PIO range does not match\n"); in hfi1_pcie_ddinit() 114 dd->kregbase1 = ioremap(addr, RCV_ARRAY); in hfi1_pcie_ddinit() 115 if (!dd->kregbase1) { in hfi1_pcie_ddinit() 116 dd_dev_err(dd, "UC mapping of kregbase1 failed\n"); in hfi1_pcie_ddinit() [all …]
|
H A D | firmware.c | 212 static int load_fabric_serdes_firmware(struct hfi1_devdata *dd, 214 static void dump_fw_version(struct hfi1_devdata *dd); 229 static int __read_8051_data(struct hfi1_devdata *dd, u32 addr, u64 *result) in __read_8051_data() argument 237 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, reg); in __read_8051_data() 239 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, in __read_8051_data() 244 while ((read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_STATUS) in __read_8051_data() 249 dd_dev_err(dd, "timeout reading 8051 data\n"); in __read_8051_data() 256 *result = read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_RD_DATA); in __read_8051_data() 265 int read_8051_data(struct hfi1_devdata *dd, u32 addr, u32 len, u64 *result) in read_8051_data() argument 271 spin_lock_irqsave(&dd->dc8051_memlock, flags); in read_8051_data() [all …]
|
H A D | init.c | 85 static int hfi1_create_kctxt(struct hfi1_devdata *dd, in hfi1_create_kctxt() argument 94 ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); in hfi1_create_kctxt() 96 dd_dev_err(dd, "Kernel receive context allocation failed\n"); in hfi1_create_kctxt() 119 rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); in hfi1_create_kctxt() 121 dd_dev_err(dd, "Kernel send context allocation failed\n"); in hfi1_create_kctxt() 132 int hfi1_create_kctxts(struct hfi1_devdata *dd) in hfi1_create_kctxts() argument 137 dd->rcd = kcalloc_node(dd->num_rcv_contexts, sizeof(*dd->rcd), in hfi1_create_kctxts() 138 GFP_KERNEL, dd->node); in hfi1_create_kctxts() 139 if (!dd->rcd) in hfi1_create_kctxts() 142 for (i = 0; i < dd->first_dyn_alloc_ctxt; ++i) { in hfi1_create_kctxts() [all …]
|
H A D | chip.c | 101 #define emulator_rev(dd) ((dd)->irev >> 8) argument 103 #define is_emulator_p(dd) ((((dd)->irev) & 0xf) == 3) argument 104 #define is_emulator_s(dd) ((((dd)->irev) & 0xf) == 4) argument 1003 static u32 chip_to_opa_lstate(struct hfi1_devdata *dd, u32 chip_lstate); 1004 static int set_physical_link_state(struct hfi1_devdata *dd, u64 state); 1005 static void read_vc_remote_phy(struct hfi1_devdata *dd, u8 *power_management, 1007 static void read_vc_remote_fabric(struct hfi1_devdata *dd, u8 *vau, u8 *z, 1009 static void read_vc_remote_link_width(struct hfi1_devdata *dd, 1011 static void read_vc_local_link_mode(struct hfi1_devdata *dd, u8 *misc_bits, 1013 static void read_remote_device_id(struct hfi1_devdata *dd, u16 *device_id, [all …]
|
H A D | msix.c | 16 int msix_initialize(struct hfi1_devdata *dd) in msix_initialize() argument 30 total = 1 + dd->num_sdma + dd->n_krcv_queues + dd->num_netdev_contexts; in msix_initialize() 35 ret = pci_alloc_irq_vectors(dd->pcidev, total, total, PCI_IRQ_MSIX); in msix_initialize() 37 dd_dev_err(dd, "pci_alloc_irq_vectors() failed: %d\n", ret); in msix_initialize() 41 entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), in msix_initialize() 44 pci_free_irq_vectors(dd->pcidev); in msix_initialize() 48 dd->msix_info.msix_entries = entries; in msix_initialize() 49 spin_lock_init(&dd->msix_info.msix_lock); in msix_initialize() 50 bitmap_zero(dd->msix_info.in_use_msix, total); in msix_initialize() 51 dd->msix_info.max_requested = total; in msix_initialize() [all …]
|
H A D | aspm.c | 23 static bool aspm_hw_l1_supported(struct hfi1_devdata *dd) in aspm_hw_l1_supported() argument 25 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_l1_supported() 35 pcie_capability_read_dword(dd->pcidev, PCI_EXP_LNKCAP, &dn); in aspm_hw_l1_supported() 42 return (!!dn || is_ax(dd)) && !!up; in aspm_hw_l1_supported() 46 static void aspm_hw_set_l1_ent_latency(struct hfi1_devdata *dd) in aspm_hw_set_l1_ent_latency() argument 51 pci_read_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, ®32); in aspm_hw_set_l1_ent_latency() 54 pci_write_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, reg32); in aspm_hw_set_l1_ent_latency() 57 static void aspm_hw_enable_l1(struct hfi1_devdata *dd) in aspm_hw_enable_l1() argument 59 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_enable_l1() 72 pcie_capability_clear_and_set_word(dd->pcidev, PCI_EXP_LNKCTL, in aspm_hw_enable_l1() [all …]
|
H A D | pio.c | 21 void __cm_reset(struct hfi1_devdata *dd, u64 sendctrl) in __cm_reset() argument 23 write_csr(dd, SEND_CTRL, sendctrl | SEND_CTRL_CM_RESET_SMASK); in __cm_reset() 26 sendctrl = read_csr(dd, SEND_CTRL); in __cm_reset() 33 void pio_send_control(struct hfi1_devdata *dd, int op) in pio_send_control() argument 41 spin_lock_irqsave(&dd->sendctrl_lock, flags); in pio_send_control() 43 reg = read_csr(dd, SEND_CTRL); in pio_send_control() 50 for (i = 0; i < ARRAY_SIZE(dd->vld); i++) in pio_send_control() 51 if (!dd->vld[i].mtu) in pio_send_control() 68 __cm_reset(dd, reg); in pio_send_control() 76 dd_dev_err(dd, "%s: invalid control %d\n", __func__, op); in pio_send_control() [all …]
|
H A D | netdev_rx.c | 24 struct hfi1_devdata *dd = rx->dd; in hfi1_netdev_setup_ctxt() local 31 ret = hfi1_create_rcvhdrq(dd, uctxt); in hfi1_netdev_setup_ctxt() 53 hfi1_rcvctrl(uctxt->dd, rcvctrl_ops, uctxt); in hfi1_netdev_setup_ctxt() 58 static int hfi1_netdev_allocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_allocate_ctxt() argument 64 if (dd->flags & HFI1_FROZEN) in hfi1_netdev_allocate_ctxt() 67 ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt); in hfi1_netdev_allocate_ctxt() 69 dd_dev_err(dd, "Unable to create ctxtdata, failing open\n"); in hfi1_netdev_allocate_ctxt() 85 dd_dev_info(dd, "created netdev context %d\n", uctxt->ctxt); in hfi1_netdev_allocate_ctxt() 91 static void hfi1_netdev_deallocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_deallocate_ctxt() argument 100 hfi1_rcvctrl(dd, HFI1_RCVCTRL_CTXT_DIS | in hfi1_netdev_deallocate_ctxt() [all …]
|
/linux/drivers/clk/ti/ |
H A D | dpll3xxx.c | 49 const struct dpll_data *dd; in _omap3_dpll_write_clken() local 52 dd = clk->dpll_data; in _omap3_dpll_write_clken() 54 v = ti_clk_ll_ops->clk_readl(&dd->control_reg); in _omap3_dpll_write_clken() 55 v &= ~dd->enable_mask; in _omap3_dpll_write_clken() 56 v |= clken_bits << __ffs(dd->enable_mask); in _omap3_dpll_write_clken() 57 ti_clk_ll_ops->clk_writel(v, &dd->control_reg); in _omap3_dpll_write_clken() 63 const struct dpll_data *dd; in _omap3_wait_dpll_status() local 68 dd = clk->dpll_data; in _omap3_wait_dpll_status() 71 state <<= __ffs(dd->idlest_mask); in _omap3_wait_dpll_status() 73 while (((ti_clk_ll_ops->clk_readl(&dd->idlest_reg) & dd->idlest_mask) in _omap3_wait_dpll_status() [all …]
|
/linux/drivers/md/ |
H A D | dm-dust.c | 78 static int dust_remove_block(struct dust_device *dd, unsigned long long block) in dust_remove_block() argument 83 spin_lock_irqsave(&dd->dust_lock, flags); in dust_remove_block() 84 bblock = dust_rb_search(&dd->badblocklist, block); in dust_remove_block() 87 if (!dd->quiet_mode) { in dust_remove_block() 91 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block() 95 rb_erase(&bblock->node, &dd->badblocklist); in dust_remove_block() 96 dd->badblock_count--; in dust_remove_block() 97 if (!dd->quiet_mode) in dust_remove_block() 100 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block() 105 static int dust_add_block(struct dust_device *dd, unsigned long long block, in dust_add_block() argument [all …]
|
/linux/sound/soc/atmel/ |
H A D | mchp-pdmc.c | 109 struct mchp_pdmc *dd; 148 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); in mchp_pdmc_sinc_order_put() 152 item = snd_soc_enum_val_to_item(e, dd->sinc_order); in mchp_pdmc_sinc_order_put() local 162 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); in mchp_pdmc_sinc_order_put() 172 if (atomic_read(&dd->busy_stream)) in mchp_pdmc_af_get() 175 if (val == dd->sinc_order) in mchp_pdmc_af_get() 178 dd->sinc_order = val; in mchp_pdmc_af_get() 187 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); in mchp_pdmc_af_put() 189 uvalue->value.integer.value[0] = !!dd->audio_filter_en; in mchp_pdmc_af_put() 198 struct mchp_pdmc *dd in mchp_pdmc_chmap_ctl_info() 100 struct mchp_pdmc *dd; global() member 138 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); mchp_pdmc_sinc_order_get() local 173 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); mchp_pdmc_af_get() local 184 struct mchp_pdmc *dd = snd_soc_component_get_drvdata(component); mchp_pdmc_af_put() local 233 struct mchp_pdmc *dd = info->dd; mchp_pdmc_chmap_ctl_get() local 275 struct mchp_pdmc *dd = info->dd; mchp_pdmc_chmap_ctl_put() local 448 struct mchp_pdmc *dd = snd_soc_dai_get_drvdata(dai); mchp_pdmc_startup() local 460 struct mchp_pdmc *dd = snd_soc_dai_get_drvdata(dai); mchp_pdmc_dai_probe() local 543 struct mchp_pdmc *dd = snd_soc_dai_get_drvdata(dai); mchp_pdmc_hw_params() local 626 mchp_pdmc_noise_filter_workaround(struct mchp_pdmc * dd) mchp_pdmc_noise_filter_workaround() argument 652 struct mchp_pdmc *dd = snd_soc_dai_get_drvdata(dai); mchp_pdmc_trigger() local 701 mchp_pdmc_add_chmap_ctls(struct snd_pcm * pcm,struct mchp_pdmc * dd) mchp_pdmc_add_chmap_ctls() argument 743 struct mchp_pdmc *dd = snd_soc_dai_get_drvdata(dai); mchp_pdmc_pcm_new() local 778 struct mchp_pdmc *dd = dev_id; mchp_pdmc_interrupt() local 869 mchp_pdmc_dt_init(struct mchp_pdmc * dd) mchp_pdmc_dt_init() argument 978 struct mchp_pdmc *dd = dev_get_drvdata(dev); mchp_pdmc_runtime_suspend() local 990 struct mchp_pdmc *dd = dev_get_drvdata(dev); mchp_pdmc_runtime_resume() local 1022 struct mchp_pdmc *dd; mchp_pdmc_probe() local 1126 struct mchp_pdmc *dd = platform_get_drvdata(pdev); mchp_pdmc_remove() local [all...] |
/linux/tools/bpf/bpftool/ |
H A D | xlated_dumper.c | 26 void kernel_syms_load(struct dump_data *dd) in kernel_syms_load() argument 38 tmp = libbpf_reallocarray(dd->sym_mapping, dd->sym_count + 1, in kernel_syms_load() 39 sizeof(*dd->sym_mapping)); in kernel_syms_load() 42 free(dd->sym_mapping); in kernel_syms_load() 43 dd->sym_mapping = NULL; in kernel_syms_load() 47 dd->sym_mapping = tmp; in kernel_syms_load() 48 sym = &dd->sym_mapping[dd->sym_count]; in kernel_syms_load() 57 dd in kernel_syms_load() 72 kernel_syms_destroy(struct dump_data * dd) kernel_syms_destroy() argument 77 kernel_syms_search(struct dump_data * dd,unsigned long key) kernel_syms_search() argument 145 print_call_pcrel(struct dump_data * dd,struct kernel_sym * sym,unsigned long address,const struct bpf_insn * insn) print_call_pcrel() argument 163 print_call_helper(struct dump_data * dd,struct kernel_sym * sym,unsigned long address) print_call_helper() argument 179 struct dump_data *dd = private_data; print_call() local 198 struct dump_data *dd = private_data; print_imm() local 218 dump_xlated_json(struct dump_data * dd,void * buf,unsigned int len,bool opcodes,bool linum) dump_xlated_json() argument 303 dump_xlated_plain(struct dump_data * dd,void * buf,unsigned int len,bool opcodes,bool linum) dump_xlated_plain() argument 367 dump_xlated_for_graph(struct dump_data * dd,void * buf_start,void * buf_end,unsigned int start_idx,bool opcodes,bool linum) dump_xlated_for_graph() argument [all...] |