| /linux/drivers/gpu/drm/bridge/analogix/ |
| H A D | analogix_dp_core.c | 40 static void analogix_dp_init_dp(struct analogix_dp_device *dp) in analogix_dp_init_dp() argument 42 analogix_dp_reset(dp); in analogix_dp_init_dp() 44 analogix_dp_swreset(dp); in analogix_dp_init_dp() 46 analogix_dp_init_analog_param(dp); in analogix_dp_init_dp() 47 analogix_dp_init_interrupt(dp); in analogix_dp_init_dp() 50 analogix_dp_enable_sw_function(dp); in analogix_dp_init_dp() 52 analogix_dp_config_interrupt(dp); in analogix_dp_init_dp() 54 analogix_dp_init_hpd(dp); in analogix_dp_init_dp() 55 analogix_dp_init_aux(dp); in analogix_dp_init_dp() 58 static int analogix_dp_detect_hpd(struct analogix_dp_device *dp) in analogix_dp_detect_hpd() argument [all …]
|
| /linux/drivers/gpu/drm/xlnx/ |
| H A D | zynqmp_dp.c | 343 struct zynqmp_dp *dp; member 416 static void zynqmp_dp_write(struct zynqmp_dp *dp, int offset, u32 val) in zynqmp_dp_write() argument 418 writel(val, dp->iomem + offset); in zynqmp_dp_write() 421 static u32 zynqmp_dp_read(struct zynqmp_dp *dp, int offset) in zynqmp_dp_read() argument 423 return readl(dp->iomem + offset); in zynqmp_dp_read() 426 static void zynqmp_dp_clr(struct zynqmp_dp *dp, int offset, u32 clr) in zynqmp_dp_clr() argument 428 zynqmp_dp_write(dp, offset, zynqmp_dp_read(dp, offset) & ~clr); in zynqmp_dp_clr() 431 static void zynqmp_dp_set(struct zynqmp_dp *dp, int offset, u32 set) in zynqmp_dp_set() argument 433 zynqmp_dp_write(dp, offset, zynqmp_dp_read(dp, offset) | set); in zynqmp_dp_set() 442 static int zynqmp_dp_reset(struct zynqmp_dp *dp, bool assert) in zynqmp_dp_reset() argument [all …]
|
| /linux/net/dsa/ |
| H A D | port.c | 33 static int dsa_port_notify(const struct dsa_port *dp, unsigned long e, void *v) in dsa_port_notify() argument 35 return dsa_tree_notify(dp->ds->dst, e, v); in dsa_port_notify() 38 static void dsa_port_notify_bridge_fdb_flush(const struct dsa_port *dp, u16 vid) in dsa_port_notify_bridge_fdb_flush() argument 40 struct net_device *brport_dev = dsa_port_to_bridge_port(dp); in dsa_port_notify_bridge_fdb_flush() 55 static void dsa_port_fast_age(const struct dsa_port *dp) in dsa_port_fast_age() argument 57 struct dsa_switch *ds = dp->ds; in dsa_port_fast_age() 62 ds->ops->port_fast_age(ds, dp->index); in dsa_port_fast_age() 65 dsa_port_notify_bridge_fdb_flush(dp, 0); in dsa_port_fast_age() 68 static int dsa_port_vlan_fast_age(const struct dsa_port *dp, u16 vid) in dsa_port_vlan_fast_age() argument 70 struct dsa_switch *ds = dp->ds; in dsa_port_vlan_fast_age() [all …]
|
| H A D | port.h | 18 bool dsa_port_supports_hwtstamp(struct dsa_port *dp); 21 int dsa_port_set_state(struct dsa_port *dp, u8 state, bool do_fast_age); 22 int dsa_port_set_mst_state(struct dsa_port *dp, 25 int dsa_port_enable_rt(struct dsa_port *dp, struct phy_device *phy); 26 int dsa_port_enable(struct dsa_port *dp, struct phy_device *phy); 27 void dsa_port_disable_rt(struct dsa_port *dp); 28 void dsa_port_disable(struct dsa_port *dp); 29 int dsa_port_bridge_join(struct dsa_port *dp, struct net_device *br, 31 void dsa_port_pre_bridge_leave(struct dsa_port *dp, struct net_device *br); 32 void dsa_port_bridge_leave(struct dsa_port *dp, struct net_device *br); [all …]
|
| H A D | switch.c | 26 struct dsa_port *dp; in dsa_switch_fastest_ageing_time() local 28 dsa_switch_for_each_port(dp, ds) in dsa_switch_fastest_ageing_time() 29 if (dp->ageing_time && dp->ageing_time < ageing_time) in dsa_switch_fastest_ageing_time() 30 ageing_time = dp->ageing_time; in dsa_switch_fastest_ageing_time() 55 static bool dsa_port_mtu_match(struct dsa_port *dp, in dsa_port_mtu_match() argument 58 return dp == info->dp || dsa_port_is_dsa(dp) || dsa_port_is_cpu(dp); in dsa_port_mtu_match() 64 struct dsa_port *dp; in dsa_switch_mtu() local 70 dsa_switch_for_each_port(dp, ds) { in dsa_switch_mtu() 71 if (dsa_port_mtu_match(dp, info)) { in dsa_switch_mtu() 72 ret = ds->ops->port_change_mtu(ds, dp->index, in dsa_switch_mtu() [all …]
|
| H A D | dsa.c | 108 struct dsa_port *dp; in dsa_tree_lag_find() local 110 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_lag_find() 111 if (dsa_port_lag_dev_get(dp) == lag_dev) in dsa_tree_lag_find() 112 return dp->lag; in dsa_tree_lag_find() 120 struct dsa_port *dp; in dsa_tree_bridge_find() local 122 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_bridge_find() 123 if (dsa_port_bridge_dev_get(dp) == br) in dsa_tree_bridge_find() 124 return dp->bridge; in dsa_tree_bridge_find() 183 struct dsa_port *dp; in dsa_switch_find() local 189 list_for_each_entry(dp, &dst->ports, list) { in dsa_switch_find() [all …]
|
| H A D | user.c | 88 struct dsa_port *dp = dsa_user_to_port(dev); in dsa_user_standalone_event_work() local 90 struct dsa_switch *ds = dp->ds; in dsa_user_standalone_event_work() 96 err = dsa_port_standalone_host_fdb_add(dp, addr, vid); in dsa_user_standalone_event_work() 100 dp->index, addr, vid, err); in dsa_user_standalone_event_work() 106 err = dsa_port_standalone_host_fdb_del(dp, addr, vid); in dsa_user_standalone_event_work() 110 dp->index, addr, vid, err); in dsa_user_standalone_event_work() 118 err = dsa_port_standalone_host_mdb_add(dp, &mdb); in dsa_user_standalone_event_work() 122 dp->index, addr, vid, err); in dsa_user_standalone_event_work() 130 err = dsa_port_standalone_host_mdb_del(dp, &mdb); in dsa_user_standalone_event_work() 134 dp->index, addr, vid, err); in dsa_user_standalone_event_work() [all …]
|
| /linux/drivers/gpu/drm/hisilicon/hibmc/dp/ |
| H A D | dp_link.c | 12 static inline int hibmc_dp_get_serdes_rate_cfg(struct hibmc_dp_dev *dp) in hibmc_dp_get_serdes_rate_cfg() argument 14 switch (dp->link.cap.link_rate) { in hibmc_dp_get_serdes_rate_cfg() 28 static int hibmc_dp_link_training_configure(struct hibmc_dp_dev *dp) in hibmc_dp_link_training_configure() argument 34 hibmc_dp_reg_write_field(dp, HIBMC_DP_PHYIF_CTRL0, HIBMC_DP_CFG_LANE_DATA_EN, in hibmc_dp_link_training_configure() 35 dp->link.cap.lanes == 0x2 ? 0x3 : 0x1); in hibmc_dp_link_training_configure() 36 hibmc_dp_reg_write_field(dp, HIBMC_DP_DPTX_GCTL0, HIBMC_DP_CFG_PHY_LANE_NUM, in hibmc_dp_link_training_configure() 37 dp->link.cap.lanes == 0x2 ? 0x1 : 0); in hibmc_dp_link_training_configure() 40 hibmc_dp_reg_write_field(dp, HIBMC_DP_VIDEO_CTRL, HIBMC_DP_CFG_STREAM_FRAME_MODE, 0x1); in hibmc_dp_link_training_configure() 43 buf[0] = dp->link.cap.link_rate; in hibmc_dp_link_training_configure() 44 buf[1] = DP_LANE_COUNT_ENHANCED_FRAME_EN | dp->link.cap.lanes; in hibmc_dp_link_training_configure() [all …]
|
| /linux/drivers/usb/typec/altmodes/ |
| H A D | displayport.c | 84 static int dp_altmode_notify(struct dp_altmode *dp) in dp_altmode_notify() argument 89 if (dp->data.conf) { in dp_altmode_notify() 90 state = get_count_order(DP_CONF_GET_PIN_ASSIGN(dp->data.conf)); in dp_altmode_notify() 96 return typec_altmode_notify(dp->alt, conf, &dp->data); in dp_altmode_notify() 99 static int dp_altmode_configure(struct dp_altmode *dp, u8 con) in dp_altmode_configure() argument 106 signal = DP_CAP_DP_SIGNALLING(dp->port->vdo) & DP_CAP_DP_SIGNALLING(dp->alt->vdo); in dp_altmode_configure() 107 if (dp->plug_prime) in dp_altmode_configure() 108 signal &= DP_CAP_DP_SIGNALLING(dp->plug_prime->vdo); in dp_altmode_configure() 117 pin_assign = DP_CAP_UFP_D_PIN_ASSIGN(dp->alt->vdo) & in dp_altmode_configure() 118 DP_CAP_DFP_D_PIN_ASSIGN(dp->port->vdo); in dp_altmode_configure() [all …]
|
| /linux/drivers/usb/typec/ucsi/ |
| H A D | displayport.c | 50 struct ucsi_dp *dp = typec_altmode_get_drvdata(alt); in ucsi_displayport_enter() local 51 struct ucsi *ucsi = dp->con->ucsi; in ucsi_displayport_enter() 57 if (!ucsi_con_mutex_lock(dp->con)) in ucsi_displayport_enter() 60 if (!dp->override && dp->initialized) { in ucsi_displayport_enter() 69 command = UCSI_GET_CURRENT_CAM | UCSI_CONNECTOR_NUMBER(dp->con->num); in ucsi_displayport_enter() 78 ret = dp->con->port_altmode[cur] == alt ? 0 : -EBUSY; in ucsi_displayport_enter() 94 dp->header = VDO(USB_TYPEC_DP_SID, 1, svdm_version, CMD_ENTER_MODE); in ucsi_displayport_enter() 95 dp->header |= VDO_OPOS(USB_TYPEC_DP_MODE); in ucsi_displayport_enter() 96 dp->header |= VDO_CMDT(CMDT_RSP_ACK); in ucsi_displayport_enter() 98 dp->vdo_data = NULL; in ucsi_displayport_enter() [all …]
|
| /linux/sound/core/seq/oss/ |
| H A D | seq_oss_init.c | 42 static int create_port(struct seq_oss_devinfo *dp); 43 static int delete_port(struct seq_oss_devinfo *dp); 44 static int alloc_seq_queue(struct seq_oss_devinfo *dp); 169 struct seq_oss_devinfo *dp; in snd_seq_oss_open() local 171 dp = kzalloc_obj(*dp); in snd_seq_oss_open() 172 if (!dp) in snd_seq_oss_open() 175 dp->cseq = system_client; in snd_seq_oss_open() 176 dp->port = -1; in snd_seq_oss_open() 177 dp->queue = -1; in snd_seq_oss_open() 184 dp->index = i; in snd_seq_oss_open() [all …]
|
| H A D | seq_oss_ioctl.c | 18 static int snd_seq_oss_synth_info_user(struct seq_oss_devinfo *dp, void __user *arg) in snd_seq_oss_synth_info_user() argument 24 if (snd_seq_oss_synth_make_info(dp, info.device, &info) < 0) in snd_seq_oss_synth_info_user() 31 static int snd_seq_oss_midi_info_user(struct seq_oss_devinfo *dp, void __user *arg) in snd_seq_oss_midi_info_user() argument 37 if (snd_seq_oss_midi_make_info(dp, info.device, &info) < 0) in snd_seq_oss_midi_info_user() 44 static int snd_seq_oss_oob_user(struct seq_oss_devinfo *dp, void __user *arg) in snd_seq_oss_oob_user() argument 52 snd_seq_oss_fill_addr(dp, &tmpev, dp->addr.client, dp->addr.port); in snd_seq_oss_oob_user() 54 if (! snd_seq_oss_process_event(dp, (union evrec *)ev, &tmpev)) { in snd_seq_oss_oob_user() 55 snd_seq_oss_dispatch(dp, &tmpev, 0, 0); in snd_seq_oss_oob_user() 61 snd_seq_oss_ioctl(struct seq_oss_devinfo *dp, unsigned int cmd, unsigned long carg) in snd_seq_oss_ioctl() argument 77 return snd_seq_oss_timer_ioctl(dp->timer, cmd, arg); in snd_seq_oss_ioctl() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_dp.c | 17 void *nfp_net_rx_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr) in nfp_net_rx_alloc_one() argument 21 if (!dp->xdp_prog) { in nfp_net_rx_alloc_one() 22 frag = netdev_alloc_frag(dp->fl_bufsz); in nfp_net_rx_alloc_one() 30 nn_dp_warn(dp, "Failed to alloc receive page frag\n"); in nfp_net_rx_alloc_one() 34 *dma_addr = nfp_net_dma_map_rx(dp, frag); in nfp_net_rx_alloc_one() 35 if (dma_mapping_error(dp->dev, *dma_addr)) { in nfp_net_rx_alloc_one() 36 nfp_net_free_frag(frag, dp->xdp_prog); in nfp_net_rx_alloc_one() 37 nn_dp_warn(dp, "Failed to map DMA RX buffer\n"); in nfp_net_rx_alloc_one() 53 nfp_net_tx_ring_init(struct nfp_net_tx_ring *tx_ring, struct nfp_net_dp *dp, in nfp_net_tx_ring_init() argument 65 tx_ring->txrwb = dp->txrwb ? &dp->txrwb[idx] : NULL; in nfp_net_tx_ring_init() [all …]
|
| H A D | nfp_net_dp.h | 9 static inline dma_addr_t nfp_net_dma_map_rx(struct nfp_net_dp *dp, void *frag) in nfp_net_dma_map_rx() argument 11 return dma_map_single_attrs(dp->dev, frag + NFP_NET_RX_BUF_HEADROOM, in nfp_net_dma_map_rx() 12 dp->fl_bufsz - NFP_NET_RX_BUF_NON_DATA, in nfp_net_dma_map_rx() 13 dp->rx_dma_dir, DMA_ATTR_SKIP_CPU_SYNC); in nfp_net_dma_map_rx() 17 nfp_net_dma_sync_dev_rx(const struct nfp_net_dp *dp, dma_addr_t dma_addr) in nfp_net_dma_sync_dev_rx() argument 19 dma_sync_single_for_device(dp->dev, dma_addr, in nfp_net_dma_sync_dev_rx() 20 dp->fl_bufsz - NFP_NET_RX_BUF_NON_DATA, in nfp_net_dma_sync_dev_rx() 21 dp->rx_dma_dir); in nfp_net_dma_sync_dev_rx() 24 static inline void nfp_net_dma_unmap_rx(struct nfp_net_dp *dp, in nfp_net_dma_unmap_rx() argument 27 dma_unmap_single_attrs(dp->dev, dma_addr, in nfp_net_dma_unmap_rx() [all …]
|
| /linux/drivers/gpu/drm/bridge/synopsys/ |
| H A D | dw-dp.c | 424 static struct dw_dp_bridge_state *dw_dp_get_bridge_state(struct dw_dp *dp) in dw_dp_get_bridge_state() argument 429 state = drm_priv_to_bridge_state(dp->bridge.base.state); in dw_dp_get_bridge_state() 440 static inline void dw_dp_phy_set_pattern(struct dw_dp *dp, u32 pattern) in dw_dp_phy_set_pattern() argument 442 regmap_update_bits(dp->regmap, DW_DP_PHYIF_CTRL, TPS_SEL, in dw_dp_phy_set_pattern() 446 static void dw_dp_phy_xmit_enable(struct dw_dp *dp, u32 lanes) in dw_dp_phy_xmit_enable() argument 462 regmap_update_bits(dp->regmap, DW_DP_PHYIF_CTRL, XMIT_ENABLE, in dw_dp_phy_xmit_enable() 466 static bool dw_dp_bandwidth_ok(struct dw_dp *dp, in dw_dp_bandwidth_ok() argument 480 static bool dw_dp_hpd_detect(struct dw_dp *dp) in dw_dp_hpd_detect() argument 484 regmap_read(dp->regmap, DW_DP_HPD_STATUS, &value); in dw_dp_hpd_detect() 510 static int dw_dp_link_parse(struct dw_dp *dp, struct drm_connector *connector) in dw_dp_link_parse() argument [all …]
|
| /linux/drivers/gpu/drm/rockchip/ |
| H A D | analogix_dp-rockchip.c | 117 static int rockchip_dp_pre_init(struct rockchip_dp_device *dp) in rockchip_dp_pre_init() argument 119 reset_control_assert(dp->rst); in rockchip_dp_pre_init() 121 reset_control_deassert(dp->rst); in rockchip_dp_pre_init() 123 reset_control_assert(dp->apbrst); in rockchip_dp_pre_init() 125 reset_control_deassert(dp->apbrst); in rockchip_dp_pre_init() 132 struct rockchip_dp_device *dp = pdata_encoder_to_dp(plat_data); in rockchip_dp_poweron() local 135 ret = clk_prepare_enable(dp->pclk); in rockchip_dp_poweron() 137 DRM_DEV_ERROR(dp->dev, "failed to enable pclk %d\n", ret); in rockchip_dp_poweron() 141 ret = rockchip_dp_pre_init(dp); in rockchip_dp_poweron() 143 DRM_DEV_ERROR(dp->dev, "failed to dp pre init %d\n", ret); in rockchip_dp_poweron() [all …]
|
| /linux/drivers/net/wireless/ath/ath12k/ |
| H A D | dp.c | 25 struct ath12k_dp *dp = ath12k_ab_to_dp(ab); in ath12k_dp_peer_cleanup() local 29 spin_lock_bh(&dp->dp_lock); in ath12k_dp_peer_cleanup() 30 peer = ath12k_dp_link_peer_find_by_vdev_and_addr(dp, vdev_id, addr); in ath12k_dp_peer_cleanup() 34 spin_unlock_bh(&dp->dp_lock); in ath12k_dp_peer_cleanup() 39 spin_unlock_bh(&dp->dp_lock); in ath12k_dp_peer_cleanup() 46 spin_unlock_bh(&dp->dp_lock); in ath12k_dp_peer_cleanup() 55 struct ath12k_dp *dp = ath12k_ab_to_dp(ab); in ath12k_dp_peer_setup() local 58 reo_dest = ar->dp.mac_id + 1; in ath12k_dp_peer_setup() 90 spin_lock_bh(&dp->dp_lock); in ath12k_dp_peer_setup() 92 peer = ath12k_dp_link_peer_find_by_vdev_and_addr(dp, vdev_id, addr); in ath12k_dp_peer_setup() [all …]
|
| H A D | dp_peer.c | 21 ath12k_dp_link_peer_find_by_vdev_and_addr(struct ath12k_dp *dp, in ath12k_dp_link_peer_find_by_vdev_and_addr() argument 26 lockdep_assert_held(&dp->dp_lock); in ath12k_dp_link_peer_find_by_vdev_and_addr() 28 list_for_each_entry(peer, &dp->peers, list) { in ath12k_dp_link_peer_find_by_vdev_and_addr() 41 ath12k_dp_link_peer_find_by_pdev_and_addr(struct ath12k_dp *dp, u8 pdev_idx, in ath12k_dp_link_peer_find_by_pdev_and_addr() argument 46 lockdep_assert_held(&dp->dp_lock); in ath12k_dp_link_peer_find_by_pdev_and_addr() 48 list_for_each_entry(peer, &dp->peers, list) { in ath12k_dp_link_peer_find_by_pdev_and_addr() 61 ath12k_dp_link_peer_find_by_addr(struct ath12k_dp *dp, const u8 *addr) in ath12k_dp_link_peer_find_by_addr() argument 63 lockdep_assert_held(&dp->dp_lock); in ath12k_dp_link_peer_find_by_addr() 65 return rhashtable_lookup_fast(dp->rhead_peer_addr, addr, in ath12k_dp_link_peer_find_by_addr() 66 dp->rhash_peer_addr_param); in ath12k_dp_link_peer_find_by_addr() [all …]
|
| /linux/drivers/gpu/drm/exynos/ |
| H A D | exynos_dp.c | 51 struct exynos_dp_device *dp = to_dp(plat_data); in exynos_dp_crtc_clock_enable() local 52 struct drm_encoder *encoder = &dp->encoder; in exynos_dp_crtc_clock_enable() 75 struct exynos_dp_device *dp = to_dp(plat_data); in exynos_dp_get_modes() local 78 if (dp->plat_data.panel) in exynos_dp_get_modes() 83 DRM_DEV_ERROR(dp->dev, in exynos_dp_get_modes() 88 drm_display_mode_from_videomode(&dp->vm, mode); in exynos_dp_get_modes() 103 struct exynos_dp_device *dp = to_dp(plat_data); in exynos_dp_bridge_attach() local 106 dp->connector = connector; in exynos_dp_bridge_attach() 109 if (dp->ptn_bridge) { in exynos_dp_bridge_attach() 110 ret = drm_bridge_attach(&dp->encoder, dp->ptn_bridge, bridge, in exynos_dp_bridge_attach() [all …]
|
| /linux/drivers/net/wireless/ath/ath12k/wifi7/ |
| H A D | dp.c | 19 static int ath12k_wifi7_dp_service_srng(struct ath12k_dp *dp, in ath12k_wifi7_dp_service_srng() argument 31 if (dp->hw_params->ring_mask->tx[grp_id]) { in ath12k_wifi7_dp_service_srng() 32 i = fls(dp->hw_params->ring_mask->tx[grp_id]) - 1; in ath12k_wifi7_dp_service_srng() 33 ath12k_wifi7_dp_tx_completion_handler(dp, i); in ath12k_wifi7_dp_service_srng() 36 if (dp->hw_params->ring_mask->rx_err[grp_id]) { in ath12k_wifi7_dp_service_srng() 37 work_done = ath12k_wifi7_dp_rx_process_err(dp, napi, budget); in ath12k_wifi7_dp_service_srng() 44 if (dp->hw_params->ring_mask->rx_wbm_rel[grp_id]) { in ath12k_wifi7_dp_service_srng() 45 work_done = ath12k_wifi7_dp_rx_process_wbm_err(dp, napi, budget); in ath12k_wifi7_dp_service_srng() 53 if (dp->hw_params->ring_mask->rx[grp_id]) { in ath12k_wifi7_dp_service_srng() 54 i = fls(dp->hw_params->ring_mask->rx[grp_id]) - 1; in ath12k_wifi7_dp_service_srng() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-io.c | 169 void (*get_page)(struct dpages *dp, 171 void (*next_page)(struct dpages *dp); 186 static void list_get_page(struct dpages *dp, in list_get_page() argument 189 unsigned int o = dp->context_u; in list_get_page() 190 struct page_list *pl = dp->context_ptr; in list_get_page() 197 static void list_next_page(struct dpages *dp) in list_next_page() argument 199 struct page_list *pl = dp->context_ptr; in list_next_page() 201 dp->context_ptr = pl->next; in list_next_page() 202 dp->context_u = 0; in list_next_page() 205 static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned int offset) in list_dp_init() argument [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| H A D | dp.c | 60 int ret = nvkm_i2c_aux_acquire(outp->dp.aux); in nvkm_dp_aux_xfer() 65 ret = nvkm_i2c_aux_xfer(outp->dp.aux, false, type, addr, data, size); in nvkm_dp_aux_xfer() 66 nvkm_i2c_aux_release(outp->dp.aux); in nvkm_dp_aux_xfer() 73 outp->dp.enabled = pu; in nvkm_dp_aux_pwr() 74 nvkm_dp_enable(outp, outp->dp.enabled); in nvkm_dp_aux_pwr() 105 ret = nvkm_rdaux(outp->dp.aux, addr, <->stat[0], 3); in nvkm_dp_train_sense() 114 ret = nvkm_rdaux(outp->dp.aux, addr, <->stat[4], 2); in nvkm_dp_train_sense() 119 ret = nvkm_rdaux(outp->dp.aux, DPCD_LS0C, <->pc2stat, 1); in nvkm_dp_train_sense() 144 for (i = 0; i < ior->dp.nr; i++) { in nvkm_dp_train_drive() 181 ior->func->dp->drive(ior, i, ocfg.pc, ocfg.dc, ocfg.pe, ocfg.tx_pu); in nvkm_dp_train_drive() [all …]
|
| /linux/arch/sparc/kernel/ |
| H A D | prom_irqtrans.c | 44 static unsigned int psycho_irq_build(struct device_node *dp, in psycho_irq_build() argument 74 static void __init psycho_irq_trans_init(struct device_node *dp) in psycho_irq_trans_init() argument 78 dp->irq_trans = prom_early_alloc(sizeof(struct of_irq_controller)); in psycho_irq_trans_init() 79 dp->irq_trans->irq_build = psycho_irq_build; in psycho_irq_trans_init() 81 regs = of_get_property(dp, "reg", NULL); in psycho_irq_trans_init() 82 dp->irq_trans->data = (void *) regs[2].phys_addr; in psycho_irq_trans_init() 183 static int sabre_device_needs_wsync(struct device_node *dp) in sabre_device_needs_wsync() argument 185 struct device_node *parent = dp->parent; in sabre_device_needs_wsync() 222 static unsigned int sabre_irq_build(struct device_node *dp, in sabre_irq_build() argument 259 regs = of_get_property(dp, "reg", NULL); in sabre_irq_build() [all …]
|
| H A D | ds.c | 135 void (*data)(struct ds_info *dp, 147 static void md_update_data(struct ds_info *dp, struct ds_cap_state *cp, 149 static void domain_shutdown_data(struct ds_info *dp, 152 static void domain_panic_data(struct ds_info *dp, 156 static void dr_cpu_data(struct ds_info *dp, 160 static void ds_pri_data(struct ds_info *dp, 163 static void ds_var_data(struct ds_info *dp, 221 static struct ds_cap_state *find_cap(struct ds_info *dp, u64 handle) in find_cap() argument 225 if (index >= dp->num_ds_states) in find_cap() 227 return &dp->ds_states[index]; in find_cap() [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_attr_inactive.c | 36 struct xfs_inode *dp, in xfs_attr3_rmt_stale() argument 53 error = xfs_bmapi_read(dp, (xfs_fileoff_t)blkno, blkcnt, in xfs_attr3_rmt_stale() 57 if (XFS_IS_CORRUPT(dp->i_mount, nmap != 1)) in xfs_attr3_rmt_stale() 65 error = xfs_attr_rmtval_stale(dp, &map, 0); in xfs_attr3_rmt_stale() 85 struct xfs_inode *dp, in xfs_attr3_leaf_inactive() argument 113 blkcnt = xfs_attr3_rmt_blocks(dp->i_mount, in xfs_attr3_leaf_inactive() 115 error = xfs_attr3_rmt_stale(dp, in xfs_attr3_leaf_inactive() 133 struct xfs_inode *dp, in xfs_attr3_node_inactive() argument 137 struct xfs_mount *mp = dp->i_mount; in xfs_attr3_node_inactive() 151 xfs_dirattr_mark_sick(dp, XFS_ATTR_FORK); in xfs_attr3_node_inactive() [all …]
|