| /linux/drivers/clk/ti/ |
| H A D | clk-dra7-atl.c | 56 struct dra7_atl_desc *cdesc; member 74 struct dra7_atl_desc *cdesc = to_atl_desc(hw); in atl_clk_enable() local 76 if (!cdesc->probed) in atl_clk_enable() 79 if (unlikely(!cdesc->valid)) in atl_clk_enable() 80 dev_warn(cdesc->cinfo->dev, "atl%d has not been configured\n", in atl_clk_enable() 81 cdesc->id); in atl_clk_enable() 82 pm_runtime_get_sync(cdesc->cinfo->dev); in atl_clk_enable() 84 atl_write(cdesc->cinfo, DRA7_ATL_ATLCR_REG(cdesc->id), in atl_clk_enable() 85 cdesc->divider - 1); in atl_clk_enable() 86 atl_write(cdesc->cinfo, DRA7_ATL_SWEN_REG(cdesc->id), DRA7_ATL_SWEN); in atl_clk_enable() [all …]
|
| /linux/arch/loongarch/mm/ |
| H A D | cache.c | 38 struct cache_desc *cdesc = current_cpu_data.cache_leaves + leaf; in flush_cache_leaf() local 40 nr_nodes = cache_private(cdesc) ? 1 : loongson_sysconf.nr_nodes; in flush_cache_leaf() 43 for (i = 0; i < cdesc->sets; i++) { in flush_cache_leaf() 44 for (j = 0; j < cdesc->ways; j++) { in flush_cache_leaf() 49 addr -= cdesc->ways; in flush_cache_leaf() 50 addr += cdesc->linesz; in flush_cache_leaf() 59 struct cache_desc *cdesc = current_cpu_data.cache_leaves; in __flush_cache_all() local 63 if (cache_inclusive(cdesc + leaf)) { in __flush_cache_all() 84 #define populate_cache_properties(cfg0, cdesc, level, leaf) \ argument 90 cdesc->flags |= CACHE_PRIVATE; \ [all …]
|
| /linux/drivers/crypto/inside-secure/ |
| H A D | safexcel_ring.c | 18 struct safexcel_command_desc *cdesc; in safexcel_init_ring_descriptors() local 48 cdesc = cdr->base; in safexcel_init_ring_descriptors() 51 cdesc->atok_lo = lower_32_bits(atok); in safexcel_init_ring_descriptors() 52 cdesc->atok_hi = upper_32_bits(atok); in safexcel_init_ring_descriptors() 53 cdesc = (void *)cdesc + cdr->offset; in safexcel_init_ring_descriptors() 187 struct safexcel_command_desc *cdesc; in safexcel_add_cdesc() local 189 cdesc = safexcel_ring_next_cwptr(priv, &priv->ring[ring_id].cdr, in safexcel_add_cdesc() 191 if (IS_ERR(cdesc)) in safexcel_add_cdesc() 192 return cdesc; in safexcel_add_cdesc() 194 cdesc->particle_size = data_len; in safexcel_add_cdesc() [all …]
|
| H A D | safexcel_hash.c | 75 static void safexcel_hash_token(struct safexcel_command_desc *cdesc, in safexcel_hash_token() argument 80 (struct safexcel_token *)cdesc->control_data.token; in safexcel_hash_token() 110 struct safexcel_command_desc *cdesc) in safexcel_context_control() argument 115 cdesc->control_data.control0 = ctx->alg; in safexcel_context_control() 116 cdesc->control_data.control1 = 0; in safexcel_context_control() 130 cdesc->control_data.control0 |= in safexcel_context_control() 137 cdesc->control_data.control0 |= in safexcel_context_control() 146 cdesc->control_data.control0 |= req->digest | in safexcel_context_control() 152 cdesc->control_data.control0 |= req->digest | in safexcel_context_control() 188 cdesc->control_data.control0 |= in safexcel_context_control() [all …]
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| H A D | ena_eth_com.c | 11 struct ena_eth_io_rx_cdesc_base *cdesc; in ena_com_get_next_rx_cdesc() local 18 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr in ena_com_get_next_rx_cdesc() 21 desc_phase = (READ_ONCE(cdesc->status) & ENA_ETH_IO_RX_CDESC_BASE_PHASE_MASK) >> in ena_com_get_next_rx_cdesc() 32 return cdesc; in ena_com_get_next_rx_cdesc() 237 struct ena_eth_io_rx_cdesc_base *cdesc; in ena_com_cdesc_rx_pkt_get() local 243 cdesc = ena_com_get_next_rx_cdesc(io_cq); in ena_com_cdesc_rx_pkt_get() 244 if (!cdesc) in ena_com_cdesc_rx_pkt_get() 246 status = READ_ONCE(cdesc->status); in ena_com_cdesc_rx_pkt_get() 255 count, io_cq->qid, cdesc->req_id); in ena_com_cdesc_rx_pkt_get() 357 struct ena_eth_io_rx_cdesc_base *cdesc) in ena_com_rx_set_flags() argument [all …]
|
| H A D | ena_eth_com.h | 204 struct ena_eth_io_tx_cdesc *cdesc; in ena_com_tx_comp_req_id_get() local 210 cdesc = (struct ena_eth_io_tx_cdesc *) in ena_com_tx_comp_req_id_get() 218 cdesc_phase = READ_ONCE(cdesc->flags) & ENA_ETH_IO_TX_CDESC_PHASE_MASK; in ena_com_tx_comp_req_id_get() 224 *req_id = READ_ONCE(cdesc->req_id); in ena_com_tx_comp_req_id_get() 227 cdesc->req_id); in ena_com_tx_comp_req_id_get()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | cacheflush.h | 12 static inline bool cache_present(struct cache_desc *cdesc) in cache_present() argument 14 return cdesc->flags & CACHE_PRESENT; in cache_present() 17 static inline bool cache_private(struct cache_desc *cdesc) in cache_private() argument 19 return cdesc->flags & CACHE_PRIVATE; in cache_private() 22 static inline bool cache_inclusive(struct cache_desc *cdesc) in cache_inclusive() argument 24 return cdesc->flags & CACHE_INCLUSIVE; in cache_inclusive()
|
| /linux/drivers/net/ethernet/mellanox/mlxfw/ |
| H A D | mlxfw_mfa2.c | 189 const struct mlxfw_mfa2_tlv_component_descriptor *cdesc; in mlxfw_mfa2_file_comp_validate() local 211 cdesc = mlxfw_mfa2_tlv_component_descriptor_get(mfa2_file, tlv); in mlxfw_mfa2_file_comp_validate() 212 if (!cdesc) { in mlxfw_mfa2_file_comp_validate() 217 pr_debug(" -- Component type %d\n", be16_to_cpu(cdesc->identifier)); in mlxfw_mfa2_file_comp_validate() 219 ((u64) be32_to_cpu(cdesc->cb_offset_h) << 32) in mlxfw_mfa2_file_comp_validate() 220 | be32_to_cpu(cdesc->cb_offset_l), be32_to_cpu(cdesc->size)); in mlxfw_mfa2_file_comp_validate()
|
| /linux/drivers/net/ethernet/altera/ |
| H A D | altera_sgdma.c | 172 struct sgdma_descrip __iomem *cdesc = &descbase[0]; in sgdma_tx_buffer() local 179 sgdma_setup_descrip(cdesc, /* current descriptor */ in sgdma_tx_buffer() 189 sgdma_async_write(priv, cdesc); in sgdma_tx_buffer() 341 struct sgdma_descrip __iomem *cdesc = &descbase[0]; in sgdma_async_read() local 352 sgdma_setup_descrip(cdesc, /* current descriptor */ in sgdma_async_read() 367 csrwr32(lower_32_bits(sgdma_rxphysaddr(priv, cdesc)), in sgdma_async_read()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_txrx.c | 2241 struct ice_tx_ctx_desc *cdesc; in ice_xmit_frame_ring() local 2245 cdesc = ICE_TX_CTX_DESC(tx_ring, i); in ice_xmit_frame_ring() 2250 cdesc->tunneling_params = cpu_to_le32(offload.cd_tunnel_params); in ice_xmit_frame_ring() 2251 cdesc->l2tag2 = cpu_to_le16(offload.cd_l2tag2); in ice_xmit_frame_ring() 2252 cdesc->gcs = cpu_to_le16(offload.cd_gcs_params); in ice_xmit_frame_ring() 2253 cdesc->qw1 = cpu_to_le64(offload.cd_qw1); in ice_xmit_frame_ring()
|