| /linux/drivers/hid/intel-ish-hid/ishtp/ |
| H A D | client.c | 23 static void ishtp_read_list_flush(struct ishtp_cl *cl) in ishtp_read_list_flush() argument 29 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush() 30 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush() 31 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush() 33 spin_lock(&cl->free_list_spinlock); in ishtp_read_list_flush() 34 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_read_list_flush() 35 spin_unlock(&cl->free_list_spinlock); in ishtp_read_list_flush() 37 spin_unlock_irqrestore(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush() 49 int ishtp_cl_flush_queues(struct ishtp_cl *cl) in ishtp_cl_flush_queues() argument 51 if (WARN_ON(!cl || !cl->dev)) in ishtp_cl_flush_queues() [all …]
|
| H A D | client-buffers.c | 19 int ishtp_cl_alloc_rx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_rx_ring() argument 21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring() 27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring() 28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring() 36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring() 38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring() 45 ishtp_cl_free_rx_ring(cl); in ishtp_cl_alloc_rx_ring() 57 int ishtp_cl_alloc_tx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_tx_ring() argument [all …]
|
| H A D | hbm.c | 55 static inline void ishtp_hbm_cl_hdr(struct ishtp_cl *cl, uint8_t hbm_cmd, in ishtp_hbm_cl_hdr() argument 63 cmd->host_addr = cl->host_client_id; in ishtp_hbm_cl_hdr() 64 cmd->fw_addr = cl->fw_client_id; in ishtp_hbm_cl_hdr() 76 static inline bool ishtp_hbm_cl_addr_equal(struct ishtp_cl *cl, void *buf) in ishtp_hbm_cl_addr_equal() argument 80 return cl->host_client_id == cmd->host_addr && in ishtp_hbm_cl_addr_equal() 81 cl->fw_client_id == cmd->fw_addr; in ishtp_hbm_cl_addr_equal() 262 struct ishtp_cl *cl) in ishtp_hbm_cl_flow_control_req() argument 270 spin_lock_irqsave(&cl->fc_spinlock, flags); in ishtp_hbm_cl_flow_control_req() 273 ishtp_hbm_cl_hdr(cl, ISHTP_FLOW_CONTROL_CMD, &flow_ctrl, len); in ishtp_hbm_cl_flow_control_req() 279 if (cl->out_flow_ctrl_creds) { in ishtp_hbm_cl_flow_control_req() [all …]
|
| /linux/net/sched/ |
| H A D | sch_hfsc.c | 178 static bool cl_in_el_or_vttree(struct hfsc_class *cl) in cl_in_el_or_vttree() argument 180 return ((cl->cl_flags & HFSC_FSC) && cl->cl_nactive) || in cl_in_el_or_vttree() 181 ((cl->cl_flags & HFSC_RSC) && !RB_EMPTY_NODE(&cl->el_node)); in cl_in_el_or_vttree() 190 eltree_insert(struct hfsc_class *cl) in eltree_insert() argument 192 struct rb_node **p = &cl->sched->eligible.rb_node; in eltree_insert() 199 if (cl->cl_e >= cl1->cl_e) in eltree_insert() 204 rb_link_node(&cl->el_node, parent, p); in eltree_insert() 205 rb_insert_color(&cl->el_node, &cl->sched->eligible); in eltree_insert() 209 eltree_remove(struct hfsc_class *cl) in eltree_remove() argument 211 if (!RB_EMPTY_NODE(&cl->el_node)) { in eltree_remove() [all …]
|
| H A D | sch_htb.c | 223 struct htb_class *cl; in htb_classify() local 234 cl = htb_find(skb->priority, sch); in htb_classify() 235 if (cl) { in htb_classify() 236 if (cl->level == 0) in htb_classify() 237 return cl; in htb_classify() 239 tcf = rcu_dereference_bh(cl->filter_list); in htb_classify() 257 cl = (void *)res.class; in htb_classify() 258 if (!cl) { in htb_classify() 261 cl = htb_find(res.classid, sch); in htb_classify() 262 if (!cl) in htb_classify() 288 htb_add_to_id_tree(struct rb_root * root,struct htb_class * cl,int prio) htb_add_to_id_tree() argument 317 htb_add_to_wait_tree(struct htb_sched * q,struct htb_class * cl,s64 delay) htb_add_to_wait_tree() argument 365 htb_add_class_to_row(struct htb_sched * q,struct htb_class * cl,int mask) htb_add_class_to_row() argument 397 htb_remove_class_from_row(struct htb_sched * q,struct htb_class * cl,int mask) htb_remove_class_from_row() argument 426 htb_activate_prios(struct htb_sched * q,struct htb_class * cl) htb_activate_prios() argument 466 htb_deactivate_prios(struct htb_sched * q,struct htb_class * cl) htb_deactivate_prios() argument 503 htb_lowater(const struct htb_class * cl) htb_lowater() argument 510 htb_hiwater(const struct htb_class * cl) htb_hiwater() argument 533 htb_class_mode(struct htb_class * cl,s64 * diff) htb_class_mode() argument 562 htb_change_class_mode(struct htb_sched * q,struct htb_class * cl,s64 * diff) htb_change_class_mode() argument 593 htb_activate(struct htb_sched * q,struct htb_class * cl) htb_activate() argument 611 htb_deactivate(struct htb_sched * q,struct htb_class * cl) htb_deactivate() argument 625 struct htb_class *cl = htb_classify(skb, sch, &ret); htb_enqueue() local 658 htb_accnt_tokens(struct htb_class * cl,int bytes,s64 diff) htb_accnt_tokens() argument 671 htb_accnt_ctokens(struct htb_class * cl,int bytes,s64 diff) htb_accnt_ctokens() argument 699 htb_charge_class(struct htb_sched * q,struct htb_class * cl,int level,struct sk_buff * skb) htb_charge_class() argument 758 struct htb_class *cl; htb_do_events() local 793 struct htb_class *cl = htb_id_find_next_upper() local 855 struct htb_class *cl; htb_lookup_leaf() local 878 struct htb_class *cl, *start; htb_dequeue_tree() local 1003 struct htb_class *cl; htb_reset() local 1245 struct htb_class *cl = (struct htb_class *)arg; htb_dump_class() local 1292 htb_offload_aggregate_stats(struct htb_sched * q,struct htb_class * cl) htb_offload_aggregate_stats() argument 1324 struct htb_class *cl = (struct htb_class *)arg; htb_dump_class_stats() local 1400 htb_offload_get_queue(struct htb_class * cl) htb_offload_get_queue() argument 1447 struct htb_class *cl = (struct htb_class *)arg; htb_graft() local 1482 struct htb_class *cl = (struct htb_class *)arg; htb_leaf() local 1488 struct htb_class *cl = (struct htb_class *)arg; htb_qlen_notify() local 1493 htb_parent_last_child(struct htb_class * cl) htb_parent_last_child() argument 1504 htb_parent_to_leaf(struct Qdisc * sch,struct htb_class * cl,struct Qdisc * new_q) htb_parent_to_leaf() argument 1540 htb_destroy_class_offload(struct Qdisc * sch,struct htb_class * cl,bool last_child,bool destroying,struct netlink_ext_ack * extack) htb_destroy_class_offload() argument 1603 htb_destroy_class(struct Qdisc * sch,struct htb_class * cl) htb_destroy_class() argument 1621 struct htb_class *cl; htb_destroy() local 1696 struct htb_class *cl = (struct htb_class *)arg; htb_delete() local 1764 struct htb_class *cl = (struct htb_class *)*arg, *parent; htb_change_class() local 2075 struct htb_class *cl = (struct htb_class *)arg; htb_tcf_block() local 2083 struct htb_class *cl = htb_find(classid, sch); htb_bind_filter() local 2101 struct htb_class *cl = (struct htb_class *)arg; htb_unbind_filter() local 2109 struct htb_class *cl; htb_walk() local [all...] |
| H A D | sch_drr.c | 38 static bool cl_is_active(struct drr_class *cl) in cl_is_active() argument 40 return !list_empty(&cl->alist); in cl_is_active() 63 struct drr_class *cl = (struct drr_class *)*arg; in drr_change_class() local 88 if (cl != NULL) { in drr_change_class() 90 err = gen_replace_estimator(&cl->bstats, NULL, in drr_change_class() 91 &cl->rate_est, in drr_change_class() 102 cl->quantum = quantum; in drr_change_class() 108 cl = kzalloc_obj(struct drr_class); in drr_change_class() 109 if (cl == NULL) in drr_change_class() 112 gnet_stats_basic_sync_init(&cl->bstats); in drr_change_class() [all …]
|
| H A D | sch_qfq.c | 205 static bool cl_is_active(struct qfq_class *cl) in cl_is_active() argument 207 return !list_empty(&cl->alist); in cl_is_active() 322 struct qfq_class *cl) in qfq_add_to_agg() argument 324 cl->agg = agg; in qfq_add_to_agg() 327 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg() 328 list_add_tail(&cl->alist, &agg->active); in qfq_add_to_agg() 330 cl && q->in_serv_agg != agg) /* agg was inactive */ in qfq_add_to_agg() 350 static void qfq_deactivate_class(struct qfq_sched *q, struct qfq_class *cl) in qfq_deactivate_class() argument 352 struct qfq_aggregate *agg = cl->agg; in qfq_deactivate_class() 355 list_del_init(&cl->alist); /* remove from RR queue of the aggregate */ in qfq_deactivate_class() [all …]
|
| H A D | sch_ets.c | 77 static bool cl_is_active(struct ets_class *cl) in cl_is_active() argument 79 return !list_empty(&cl->alist); in cl_is_active() 104 static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl) in ets_class_id() argument 107 int band = cl - q->classes; in ets_class_id() 201 static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl) in ets_class_is_strict() argument 203 unsigned int band = cl - q->classes; in ets_class_is_strict() 212 struct ets_class *cl = ets_class_from_arg(sch, *arg); in ets_class_change() local 222 if (!cl) { in ets_class_change() 240 if (ets_class_is_strict(q, cl)) { in ets_class_change() 251 cl->quantum = quantum; in ets_class_change() [all …]
|
| /linux/drivers/misc/mei/ |
| H A D | client.c | 319 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue() 331 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue() 332 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue() 345 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument 350 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp() 367 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument 379 cb->cl = cl; in mei_io_cb_init() 395 const struct mei_cl *cl) in mei_io_list_flush_cl() argument 400 if (cl == cb->cl) { in mei_io_list_flush_cl() 416 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument [all …]
|
| H A D | main.c | 50 struct mei_cl *cl; in mei_open() local 68 cl = mei_cl_alloc_linked(dev); in mei_open() 69 if (IS_ERR(cl)) { in mei_open() 70 err = PTR_ERR(cl); in mei_open() 74 cl->fp = file; in mei_open() 75 file->private_data = cl; in mei_open() 94 static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl, in mei_cl_vtag_remove_by_fp() argument 99 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) { in mei_cl_vtag_remove_by_fp() 118 struct mei_cl *cl = file->private_data; in mei_release() local 122 if (WARN_ON(!cl || !cl->dev)) in mei_release() [all …]
|
| H A D | interrupt.c | 32 struct mei_cl *cl; in mei_irq_compl_handler() local 35 cl = cb->cl; in mei_irq_compl_handler() 38 cl_dbg(dev, cl, "completing call back.\n"); in mei_irq_compl_handler() 39 mei_cl_complete(cl, cb); in mei_irq_compl_handler() 52 static inline int mei_cl_hbm_equal(struct mei_cl *cl, in mei_cl_hbm_equal() argument 55 return mei_cl_host_addr(cl) == mei_hdr->host_addr && in mei_cl_hbm_equal() 56 mei_cl_me_id(cl) == mei_hdr->me_addr; in mei_cl_hbm_equal() 93 static int mei_cl_irq_read_msg(struct mei_cl *cl, in mei_cl_irq_read_msg() argument 98 struct mei_device *dev = cl->dev; in mei_cl_irq_read_msg() 115 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg() [all …]
|
| H A D | bus.c | 35 ssize_t __mei_cl_send(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send() argument 38 return __mei_cl_send_timeout(cl, buf, length, vtag, mode, MAX_SCHEDULE_TIMEOUT); in __mei_cl_send() 55 ssize_t __mei_cl_send_timeout(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send_timeout() argument 62 if (WARN_ON(!cl || !cl->dev)) in __mei_cl_send_timeout() 65 bus = cl->dev; in __mei_cl_send_timeout() 74 if (!mei_cl_is_connected(cl)) { in __mei_cl_send_timeout() 80 if (!mei_me_cl_is_active(cl->me_cl)) { in __mei_cl_send_timeout() 87 rets = mei_cl_vt_support_check(cl); in __mei_cl_send_timeout() 92 if (length > mei_cl_mtu(cl)) { in __mei_cl_send_timeout() 97 while (cl->tx_cb_queued >= bus->tx_queue_limit) { in __mei_cl_send_timeout() [all …]
|
| H A D | hbm.c | 150 void mei_hbm_cl_hdr(struct mei_cl *cl, u8 hbm_cmd, void *buf, size_t len) in mei_hbm_cl_hdr() argument 157 cmd->host_addr = mei_cl_host_addr(cl); in mei_hbm_cl_hdr() 158 cmd->me_addr = mei_cl_me_id(cl); in mei_hbm_cl_hdr() 172 static inline int mei_hbm_cl_write(struct mei_device *dev, struct mei_cl *cl, in mei_hbm_cl_write() argument 178 mei_hbm_cl_hdr(cl, hbm_cmd, buf, len); in mei_hbm_cl_write() 193 bool mei_hbm_cl_addr_equal(struct mei_cl *cl, struct mei_hbm_cl_cmd *cmd) in mei_hbm_cl_addr_equal() argument 195 return mei_cl_host_addr(cl) == cmd->host_addr && in mei_hbm_cl_addr_equal() 196 mei_cl_me_id(cl) == cmd->me_addr; in mei_hbm_cl_addr_equal() 211 struct mei_cl *cl; in mei_hbm_cl_find_by_cmd() local 213 list_for_each_entry(cl, &dev->file_list, link) in mei_hbm_cl_find_by_cmd() [all …]
|
| /linux/sound/soc/intel/avs/ |
| H A D | cldma.c | 60 void hda_cldma_fill(struct hda_cldma *cl) in hda_cldma_fill() argument 64 if (cl->remaining > cl->buffer_size) in hda_cldma_fill() 65 size = cl->buffer_size; in hda_cldma_fill() 67 size = cl->remaining; in hda_cldma_fill() 69 offset = snd_hdac_stream_readl(cl, CL_SD_SPIB); in hda_cldma_fill() 70 if (offset + size > cl->buffer_size) { in hda_cldma_fill() 73 ss = cl->buffer_size - offset; in hda_cldma_fill() 74 memcpy(cl->dmab_data.area + offset, cl->position, ss); in hda_cldma_fill() 77 cl->position += ss; in hda_cldma_fill() 78 cl->remaining -= ss; in hda_cldma_fill() [all …]
|
| H A D | cldma.h | 18 void hda_cldma_fill(struct hda_cldma *cl); 19 void hda_cldma_transfer(struct hda_cldma *cl, unsigned long start_delay); 21 int hda_cldma_start(struct hda_cldma *cl); 22 int hda_cldma_stop(struct hda_cldma *cl); 23 int hda_cldma_reset(struct hda_cldma *cl); 25 void hda_cldma_set_data(struct hda_cldma *cl, void *data, unsigned int size); 26 void hda_cldma_setup(struct hda_cldma *cl); 27 void hda_cldma_interrupt(struct hda_cldma *cl); 28 int hda_cldma_init(struct hda_cldma *cl, struct hdac_bus *bus, void __iomem *dsp_ba, 30 void hda_cldma_free(struct hda_cldma *cl);
|
| /linux/include/linux/ |
| H A D | closure.h | 171 void closure_sub(struct closure *cl, int v); 172 void closure_put(struct closure *cl); 174 bool closure_wait(struct closure_waitlist *list, struct closure *cl); 175 void __closure_sync(struct closure *cl); 177 static inline unsigned closure_nr_remaining(struct closure *cl) in closure_nr_remaining() argument 179 return atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK; in closure_nr_remaining() 188 static inline void closure_sync(struct closure *cl) in closure_sync() argument 191 BUG_ON(closure_nr_remaining(cl) != 1 && !cl->closure_get_happened); in closure_sync() 194 if (cl->closure_get_happened) in closure_sync() 195 __closure_sync(cl); in closure_sync() [all …]
|
| /linux/lib/ |
| H A D | closure.c | 30 static inline void closure_put_after_sub(struct closure *cl, int flags) in closure_put_after_sub() argument 37 cl->closure_get_happened = false; in closure_put_after_sub() 39 if (cl->fn && !(flags & CLOSURE_DESTRUCTOR)) { in closure_put_after_sub() 40 atomic_set(&cl->remaining, in closure_put_after_sub() 42 closure_queue(cl); in closure_put_after_sub() 44 struct closure *parent = cl->parent; in closure_put_after_sub() 45 closure_fn *destructor = cl->fn; in closure_put_after_sub() 47 closure_debug_destroy(cl); in closure_put_after_sub() 50 destructor(&cl->work); in closure_put_after_sub() 59 void closure_sub(struct closure *cl, int v) in closure_sub() argument [all …]
|
| /linux/drivers/hsi/ |
| H A D | hsi_core.c | 60 struct hsi_client *cl = to_hsi_client(dev); in hsi_client_release() local 62 kfree(cl->tx_cfg.channels); in hsi_client_release() 63 kfree(cl->rx_cfg.channels); in hsi_client_release() 64 kfree(cl); in hsi_client_release() 70 struct hsi_client *cl; in hsi_new_client() local 73 cl = kzalloc_obj(*cl); in hsi_new_client() 74 if (!cl) in hsi_new_client() 77 cl->tx_cfg = info->tx_cfg; in hsi_new_client() 78 if (cl in hsi_new_client() 200 struct hsi_client *cl; hsi_add_client_from_dt() local 577 hsi_async(struct hsi_client * cl,struct hsi_msg * msg) hsi_async() argument 598 hsi_claim_port(struct hsi_client * cl,unsigned int share) hsi_claim_port() argument 626 hsi_release_port(struct hsi_client * cl) hsi_release_port() argument 647 struct hsi_client *cl = container_of(nb, struct hsi_client, nb); hsi_event_notifier_call() local 666 hsi_register_port_event(struct hsi_client * cl,void (* handler)(struct hsi_client *,unsigned long)) hsi_register_port_event() argument 691 hsi_unregister_port_event(struct hsi_client * cl) hsi_unregister_port_event() argument 738 hsi_get_channel_id_by_name(struct hsi_client * cl,char * name) hsi_get_channel_id_by_name() argument [all...] |
| /linux/drivers/hsi/clients/ |
| H A D | ssi_protocol.c | 139 struct hsi_client *cl; member 217 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); in ssip_release_cmd() 219 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); in ssip_release_cmd() 331 if (slave->device.parent == ssi->cl->device.parent) { in ssip_slave_get_master() 332 master = ssi->cl; in ssip_slave_get_master() 385 static void ssip_reset(struct hsi_client *cl) in ssip_reset() argument 387 struct ssi_protocol *ssi = hsi_client_drvdata(cl); in ssip_reset() 393 hsi_flush(cl); in ssip_reset() 396 hsi_stop_tx(cl); in ssip_reset() 399 ssi_waketest(cl, 0); /* FIXME: To be removed */ in ssip_reset() [all …]
|
| H A D | hsi_char.c | 85 struct hsi_client *cl; member 105 struct hsi_client *cl; member 227 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_rx_completed() 247 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_tx_completed() 267 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_req_destructor() 275 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_received() 294 hsi_flush(msg->cl); in hsc_break_received() 295 ret = hsi_async_read(msg->cl, msg); in hsc_break_received() 300 static int hsc_break_request(struct hsi_client *cl) in hsc_break_request() argument 302 struct hsc_client_data *cl_data = hsi_client_drvdata(cl); in hsc_break_request() [all …]
|
| /linux/drivers/clk/ |
| H A D | clkdev.c | 38 struct clk_lookup *p, *cl = NULL; in clk_find() local 62 cl = p; in clk_find() 69 return cl; in clk_find() 74 struct clk_lookup *cl; in clk_find_hw() local 78 cl = clk_find(dev_id, con_id); in clk_find_hw() 79 if (cl) in clk_find_hw() 80 hw = cl->clk_hw; in clk_find_hw() 121 static void __clkdev_add(struct clk_lookup *cl) in __clkdev_add() argument 124 list_add_tail(&cl->node, &clocks); in __clkdev_add() 128 void clkdev_add(struct clk_lookup *cl) in clkdev_add() argument [all …]
|
| /linux/include/linux/hsi/ |
| H A D | hsi.h | 138 static inline void hsi_client_set_drvdata(struct hsi_client *cl, void *data) in hsi_client_set_drvdata() argument 140 dev_set_drvdata(&cl->device, data); in hsi_client_set_drvdata() 143 static inline void *hsi_client_drvdata(struct hsi_client *cl) in hsi_client_drvdata() argument 145 return dev_get_drvdata(&cl->device); in hsi_client_drvdata() 148 int hsi_register_port_event(struct hsi_client *cl, 150 int hsi_unregister_port_event(struct hsi_client *cl); 173 * @cl: HSI device client that issues the transfer 187 struct hsi_client *cl; member 230 int (*setup)(struct hsi_client *cl); 231 int (*flush)(struct hsi_client *cl); 240 hsi_get_port(cl) global() argument 246 hsi_port_claimed(struct hsi_client * cl) hsi_port_claimed() argument 329 hsi_id(struct hsi_client * cl) hsi_id() argument 340 hsi_port_id(struct hsi_client * cl) hsi_port_id() argument 354 hsi_setup(struct hsi_client * cl) hsi_setup() argument 370 hsi_flush(struct hsi_client * cl) hsi_flush() argument 384 hsi_async_read(struct hsi_client * cl,struct hsi_msg * msg) hsi_async_read() argument 397 hsi_async_write(struct hsi_client * cl,struct hsi_msg * msg) hsi_async_write() argument 409 hsi_start_tx(struct hsi_client * cl) hsi_start_tx() argument 422 hsi_stop_tx(struct hsi_client * cl) hsi_stop_tx() argument [all...] |
| /linux/drivers/md/bcache/ |
| H A D | request.c | 60 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK() 67 op->flush_journal ? cl : NULL); in CLOSURE_CALLBACK() 82 continue_at(cl, bch_data_insert_start, op->wq); in CLOSURE_CALLBACK() 87 closure_return(cl); in CLOSURE_CALLBACK() 108 static void bch_data_invalidate(struct closure *cl) in bch_data_invalidate() argument 110 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_invalidate() 136 continue_at(cl, bch_data_insert_keys, op->wq); in bch_data_invalidate() 141 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK() 166 bch_data_insert_keys(&cl->work); in CLOSURE_CALLBACK() 171 struct closure *cl = bio->bi_private; in bch_data_insert_endio() local [all …]
|
| /linux/include/linux/soc/samsung/ |
| H A D | exynos-regs-pmu.h | 718 #define GS101_CLUSTER_CPU_OFFSET(cl, cpu) ((cl) + ((cpu) * 0x80)) argument 719 #define GS101_CLUSTER_CPU_CONFIGURATION(cl, cpu) \ argument 720 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x00) 721 #define GS101_CLUSTER_CPU_STATUS(cl, cpu) \ argument 722 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x04) 723 #define GS101_CLUSTER_CPU_STATES(cl, cpu) \ argument 724 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x08) 725 #define GS101_CLUSTER_CPU_OPTION(cl, cpu) \ argument 726 (GS101_CLUSTER_CPU_OFFSET(cl, cp 727 GS101_CLUSTER_CPU_OUT(cl,cpu) global() argument 729 GS101_CLUSTER_CPU_IN(cl,cpu) global() argument 731 GS101_CLUSTER_CPU_INT_IN(cl,cpu) global() argument 733 GS101_CLUSTER_CPU_INT_EN(cl,cpu) global() argument 735 GS101_CLUSTER_CPU_INT_TYPE(cl,cpu) global() argument 737 GS101_CLUSTER_CPU_INT_DIR(cl,cpu) global() argument 740 GS101_CLUSTER_NONCPU_OFFSET(cl) global() argument 741 GS101_CLUSTER_NONCPU_CONFIGURATION(cl) global() argument 743 GS101_CLUSTER_NONCPU_STATUS(cl) global() argument 745 GS101_CLUSTER_NONCPU_STATES(cl) global() argument 747 GS101_CLUSTER_NONCPU_OPTION(cl) global() argument 749 GS101_CLUSTER_NONCPU_OUT(cl) global() argument 751 GS101_CLUSTER_NONCPU_IN(cl) global() argument 753 GS101_CLUSTER_NONCPU_INT_IN(cl) global() argument 755 GS101_CLUSTER_NONCPU_INT_EN(cl) global() argument 757 GS101_CLUSTER_NONCPU_INT_TYPE(cl) global() argument 759 GS101_CLUSTER_NONCPU_INT_DIR(cl) global() argument 761 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_OUT(cl) global() argument 763 GS101_CLUSTER_NONCPU_DUALRAIL_POS_OUT(cl) global() argument 765 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_IN(cl) global() argument [all...] |
| /linux/arch/x86/math-emu/ |
| H A D | wm_shrx.S | 50 shrd %cl,%ebx,%eax 51 shrd %cl,%edx,%ebx 52 shr %cl,%edx 64 subb $32,%cl 67 shrd %cl,%edx,%eax 68 shr %cl,%edx 79 subb $64,%cl 81 shr %cl,%eax 131 subb $32,%cl 135 shrd %cl,%eax,%ebx [all …]
|