Home
last modified time | relevance | path

Searched refs:cdesc (Results 1 – 10 of 10) sorted by relevance

/linux/drivers/clk/ti/
H A Dclk-dra7-atl.c56 struct dra7_atl_desc *cdesc; member
74 struct dra7_atl_desc *cdesc = to_atl_desc(hw); in atl_clk_enable() local
76 if (!cdesc->probed) in atl_clk_enable()
79 if (unlikely(!cdesc->valid)) in atl_clk_enable()
80 dev_warn(cdesc->cinfo->dev, "atl%d has not been configured\n", in atl_clk_enable()
81 cdesc->id); in atl_clk_enable()
82 pm_runtime_get_sync(cdesc->cinfo->dev); in atl_clk_enable()
84 atl_write(cdesc->cinfo, DRA7_ATL_ATLCR_REG(cdesc->id), in atl_clk_enable()
85 cdesc->divider - 1); in atl_clk_enable()
86 atl_write(cdesc->cinfo, DRA7_ATL_SWEN_REG(cdesc->id), DRA7_ATL_SWEN); in atl_clk_enable()
[all …]
/linux/arch/loongarch/mm/
H A Dcache.c38 struct cache_desc *cdesc = current_cpu_data.cache_leaves + leaf; in flush_cache_leaf() local
40 nr_nodes = cache_private(cdesc) ? 1 : loongson_sysconf.nr_nodes; in flush_cache_leaf()
43 for (i = 0; i < cdesc->sets; i++) { in flush_cache_leaf()
44 for (j = 0; j < cdesc->ways; j++) { in flush_cache_leaf()
49 addr -= cdesc->ways; in flush_cache_leaf()
50 addr += cdesc->linesz; in flush_cache_leaf()
59 struct cache_desc *cdesc = current_cpu_data.cache_leaves; in __flush_cache_all() local
63 if (cache_inclusive(cdesc + leaf)) { in __flush_cache_all()
84 #define populate_cache_properties(cfg0, cdesc, level, leaf) \ argument
90 cdesc->flags |= CACHE_PRIVATE; \
[all …]
/linux/drivers/crypto/inside-secure/
H A Dsafexcel_ring.c18 struct safexcel_command_desc *cdesc; in safexcel_init_ring_descriptors() local
48 cdesc = cdr->base; in safexcel_init_ring_descriptors()
51 cdesc->atok_lo = lower_32_bits(atok); in safexcel_init_ring_descriptors()
52 cdesc->atok_hi = upper_32_bits(atok); in safexcel_init_ring_descriptors()
53 cdesc = (void *)cdesc + cdr->offset; in safexcel_init_ring_descriptors()
187 struct safexcel_command_desc *cdesc; in safexcel_add_cdesc() local
189 cdesc = safexcel_ring_next_cwptr(priv, &priv->ring[ring_id].cdr, in safexcel_add_cdesc()
191 if (IS_ERR(cdesc)) in safexcel_add_cdesc()
192 return cdesc; in safexcel_add_cdesc()
194 cdesc->particle_size = data_len; in safexcel_add_cdesc()
[all …]
H A Dsafexcel_hash.c75 static void safexcel_hash_token(struct safexcel_command_desc *cdesc, in safexcel_hash_token() argument
80 (struct safexcel_token *)cdesc->control_data.token; in safexcel_hash_token()
110 struct safexcel_command_desc *cdesc) in safexcel_context_control() argument
115 cdesc->control_data.control0 = ctx->alg; in safexcel_context_control()
116 cdesc->control_data.control1 = 0; in safexcel_context_control()
130 cdesc->control_data.control0 |= in safexcel_context_control()
137 cdesc->control_data.control0 |= in safexcel_context_control()
146 cdesc->control_data.control0 |= req->digest | in safexcel_context_control()
152 cdesc->control_data.control0 |= req->digest | in safexcel_context_control()
188 cdesc->control_data.control0 |= in safexcel_context_control()
[all …]
/linux/drivers/net/ethernet/amazon/ena/
H A Dena_eth_com.c11 struct ena_eth_io_rx_cdesc_base *cdesc; in ena_com_get_next_rx_cdesc() local
18 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr in ena_com_get_next_rx_cdesc()
21 desc_phase = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_PHASE_MASK) >> in ena_com_get_next_rx_cdesc()
32 return cdesc; in ena_com_get_next_rx_cdesc()
237 struct ena_eth_io_rx_cdesc_base *cdesc; in ena_com_cdesc_rx_pkt_get() local
243 cdesc = ena_com_get_next_rx_cdesc(io_cq); in ena_com_cdesc_rx_pkt_get()
244 if (!cdesc) in ena_com_cdesc_rx_pkt_get()
246 status = READ_ONCE(cdesc->status); in ena_com_cdesc_rx_pkt_get()
255 count, io_cq->qid, cdesc->req_id); in ena_com_cdesc_rx_pkt_get()
357 struct ena_eth_io_rx_cdesc_base *cdesc) in ena_com_rx_set_flags() argument
[all …]
H A Dena_eth_com.h204 struct ena_eth_io_tx_cdesc *cdesc; in ena_com_tx_comp_req_id_get() local
210 cdesc = (struct ena_eth_io_tx_cdesc *) in ena_com_tx_comp_req_id_get()
218 cdesc_phase = READ_ONCE(cdesc->flags) & ENA_ETH_IO_TX_CDESC_PHASE_MASK; in ena_com_tx_comp_req_id_get()
224 *req_id = READ_ONCE(cdesc->req_id); in ena_com_tx_comp_req_id_get()
227 cdesc->req_id); in ena_com_tx_comp_req_id_get()
/linux/arch/loongarch/include/asm/
H A Dcacheflush.h12 static inline bool cache_present(struct cache_desc *cdesc) in cache_present() argument
14 return cdesc->flags & CACHE_PRESENT; in cache_present()
17 static inline bool cache_private(struct cache_desc *cdesc) in cache_private() argument
19 return cdesc->flags & CACHE_PRIVATE; in cache_private()
22 static inline bool cache_inclusive(struct cache_desc *cdesc) in cache_inclusive() argument
24 return cdesc->flags & CACHE_INCLUSIVE; in cache_inclusive()
/linux/drivers/net/ethernet/mellanox/mlxfw/
H A Dmlxfw_mfa2.c189 const struct mlxfw_mfa2_tlv_component_descriptor *cdesc; in mlxfw_mfa2_file_comp_validate() local
211 cdesc = mlxfw_mfa2_tlv_component_descriptor_get(mfa2_file, tlv); in mlxfw_mfa2_file_comp_validate()
212 if (!cdesc) { in mlxfw_mfa2_file_comp_validate()
217 pr_debug(" -- Component type %d\n", be16_to_cpu(cdesc->identifier)); in mlxfw_mfa2_file_comp_validate()
219 ((u64) be32_to_cpu(cdesc->cb_offset_h) << 32) in mlxfw_mfa2_file_comp_validate()
220 | be32_to_cpu(cdesc->cb_offset_l), be32_to_cpu(cdesc->size)); in mlxfw_mfa2_file_comp_validate()
/linux/drivers/net/ethernet/altera/
H A Daltera_sgdma.c172 struct sgdma_descrip __iomem *cdesc = &descbase[0]; in sgdma_tx_buffer() local
179 sgdma_setup_descrip(cdesc, /* current descriptor */ in sgdma_tx_buffer()
189 sgdma_async_write(priv, cdesc); in sgdma_tx_buffer()
341 struct sgdma_descrip __iomem *cdesc = &descbase[0]; in sgdma_async_read() local
352 sgdma_setup_descrip(cdesc, /* current descriptor */ in sgdma_async_read()
367 csrwr32(lower_32_bits(sgdma_rxphysaddr(priv, cdesc)), in sgdma_async_read()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_txrx.c2241 struct ice_tx_ctx_desc *cdesc; in ice_xmit_frame_ring() local
2245 cdesc = ICE_TX_CTX_DESC(tx_ring, i); in ice_xmit_frame_ring()
2250 cdesc->tunneling_params = cpu_to_le32(offload.cd_tunnel_params); in ice_xmit_frame_ring()
2251 cdesc->l2tag2 = cpu_to_le16(offload.cd_l2tag2); in ice_xmit_frame_ring()
2252 cdesc->gcs = cpu_to_le16(offload.cd_gcs_params); in ice_xmit_frame_ring()
2253 cdesc->qw1 = cpu_to_le64(offload.cd_qw1); in ice_xmit_frame_ring()