Home
last modified time | relevance | path

Searched refs:cl (Results 1 – 25 of 312) sorted by relevance

12345678910>>...13

/linux/drivers/hid/intel-ish-hid/ishtp/
H A Dclient.c23 static void ishtp_read_list_flush(struct ishtp_cl *cl) in ishtp_read_list_flush() argument
29 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush()
30 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush()
31 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush()
33 spin_lock(&cl->free_list_spinlock); in ishtp_read_list_flush()
34 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_read_list_flush()
35 spin_unlock(&cl->free_list_spinlock); in ishtp_read_list_flush()
37 spin_unlock_irqrestore(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush()
49 int ishtp_cl_flush_queues(struct ishtp_cl *cl) in ishtp_cl_flush_queues() argument
51 if (WARN_ON(!cl || !cl->dev)) in ishtp_cl_flush_queues()
[all …]
H A Dclient-buffers.c19 int ishtp_cl_alloc_rx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_rx_ring() argument
21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring()
27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring()
28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring()
36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring()
45 ishtp_cl_free_rx_ring(cl); in ishtp_cl_alloc_rx_ring()
57 int ishtp_cl_alloc_tx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_tx_ring() argument
[all …]
H A Dhbm.c55 static inline void ishtp_hbm_cl_hdr(struct ishtp_cl *cl, uint8_t hbm_cmd, in ishtp_hbm_cl_hdr() argument
63 cmd->host_addr = cl->host_client_id; in ishtp_hbm_cl_hdr()
64 cmd->fw_addr = cl->fw_client_id; in ishtp_hbm_cl_hdr()
76 static inline bool ishtp_hbm_cl_addr_equal(struct ishtp_cl *cl, void *buf) in ishtp_hbm_cl_addr_equal() argument
80 return cl->host_client_id == cmd->host_addr && in ishtp_hbm_cl_addr_equal()
81 cl->fw_client_id == cmd->fw_addr; in ishtp_hbm_cl_addr_equal()
262 struct ishtp_cl *cl) in ishtp_hbm_cl_flow_control_req() argument
270 spin_lock_irqsave(&cl->fc_spinlock, flags); in ishtp_hbm_cl_flow_control_req()
273 ishtp_hbm_cl_hdr(cl, ISHTP_FLOW_CONTROL_CMD, &flow_ctrl, len); in ishtp_hbm_cl_flow_control_req()
279 if (cl->out_flow_ctrl_creds) { in ishtp_hbm_cl_flow_control_req()
[all …]
/linux/net/sched/
H A Dsch_hfsc.c178 static bool cl_in_el_or_vttree(struct hfsc_class *cl) in cl_in_el_or_vttree() argument
180 return ((cl->cl_flags & HFSC_FSC) && cl->cl_nactive) || in cl_in_el_or_vttree()
181 ((cl->cl_flags & HFSC_RSC) && !RB_EMPTY_NODE(&cl->el_node)); in cl_in_el_or_vttree()
190 eltree_insert(struct hfsc_class *cl) in eltree_insert() argument
192 struct rb_node **p = &cl->sched->eligible.rb_node; in eltree_insert()
199 if (cl->cl_e >= cl1->cl_e) in eltree_insert()
204 rb_link_node(&cl->el_node, parent, p); in eltree_insert()
205 rb_insert_color(&cl in eltree_insert()
209 eltree_remove(struct hfsc_class * cl) eltree_remove() argument
218 eltree_update(struct hfsc_class * cl) eltree_update() argument
228 struct hfsc_class *p, *cl = NULL; eltree_get_mindl() local
258 vttree_insert(struct hfsc_class * cl) vttree_insert() argument
277 vttree_remove(struct hfsc_class * cl) vttree_remove() argument
283 vttree_update(struct hfsc_class * cl) vttree_update() argument
290 vttree_firstfit(struct hfsc_class * cl,u64 cur_time) vttree_firstfit() argument
307 vttree_get_minvt(struct hfsc_class * cl,u64 cur_time) vttree_get_minvt() argument
327 cftree_insert(struct hfsc_class * cl) cftree_insert() argument
346 cftree_remove(struct hfsc_class * cl) cftree_remove() argument
352 cftree_update(struct hfsc_class * cl) cftree_update() argument
617 init_ed(struct hfsc_class * cl,unsigned int next_len) init_ed() argument
643 update_ed(struct hfsc_class * cl,unsigned int next_len) update_ed() argument
652 update_d(struct hfsc_class * cl,unsigned int next_len) update_d() argument
658 update_cfmin(struct hfsc_class * cl) update_cfmin() argument
672 init_vf(struct hfsc_class * cl,unsigned int len) init_vf() argument
751 update_vf(struct hfsc_class * cl,unsigned int len,u64 cur_time) update_vf() argument
839 hfsc_adjust_levels(struct hfsc_class * cl) hfsc_adjust_levels() argument
867 hfsc_change_rsc(struct hfsc_class * cl,struct tc_service_curve * rsc,u64 cur_time) hfsc_change_rsc() argument
881 hfsc_change_fsc(struct hfsc_class * cl,struct tc_service_curve * fsc) hfsc_change_fsc() argument
889 hfsc_change_usc(struct hfsc_class * cl,struct tc_service_curve * usc,u64 cur_time) hfsc_change_usc() argument
898 hfsc_upgrade_rt(struct hfsc_class * cl) hfsc_upgrade_rt() argument
917 struct hfsc_class *cl = (struct hfsc_class *)*arg; hfsc_change_class() local
1091 hfsc_destroy_class(struct Qdisc * sch,struct hfsc_class * cl) hfsc_destroy_class() argument
1107 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_delete_class() local
1133 struct hfsc_class *head, *cl; hfsc_classify() local
1188 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_graft_class() local
1206 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_class_leaf() local
1217 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_qlen_notify() local
1238 struct hfsc_class *cl = hfsc_find_class(classid, sch); hfsc_bind_tcf() local
1252 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_unbind_tcf() local
1261 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_tcf_block() local
1287 hfsc_dump_curves(struct sk_buff * skb,struct hfsc_class * cl) hfsc_dump_curves() argument
1311 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_dump_class() local
1336 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_dump_class_stats() local
1360 struct hfsc_class *cl; hfsc_walk() local
1379 struct hfsc_class *cl; hfsc_schedule_watchdog() local
1453 hfsc_reset_class(struct hfsc_class * cl) hfsc_reset_class() argument
1486 struct hfsc_class *cl; hfsc_reset_qdisc() local
1502 struct hfsc_class *cl; hfsc_destroy_qdisc() local
1541 struct hfsc_class *cl; hfsc_enqueue() local
1588 struct hfsc_class *cl; hfsc_dequeue() local
[all...]
H A Dsch_htb.c223 struct htb_class *cl; in htb_classify() local
234 cl = htb_find(skb->priority, sch); in htb_classify()
235 if (cl) { in htb_classify()
236 if (cl->level == 0) in htb_classify()
237 return cl; in htb_classify()
239 tcf = rcu_dereference_bh(cl->filter_list); in htb_classify()
257 cl = (void *)res.class; in htb_classify()
258 if (!cl) { in htb_classify()
261 cl = htb_find(res.classid, sch); in htb_classify()
262 if (!cl) in htb_classify()
[all …]
H A Dsch_drr.c38 static bool cl_is_active(struct drr_class *cl) in cl_is_active() argument
40 return !list_empty(&cl->alist); in cl_is_active()
63 struct drr_class *cl = (struct drr_class *)*arg; in drr_change_class() local
88 if (cl != NULL) { in drr_change_class()
90 err = gen_replace_estimator(&cl->bstats, NULL, in drr_change_class()
91 &cl->rate_est, in drr_change_class()
102 cl->quantum = quantum; in drr_change_class()
108 cl = kzalloc_obj(struct drr_class); in drr_change_class()
109 if (cl == NULL) in drr_change_class()
112 gnet_stats_basic_sync_init(&cl->bstats); in drr_change_class()
[all …]
H A Dsch_qfq.c205 static bool cl_is_active(struct qfq_class *cl) in cl_is_active() argument
207 return !list_empty(&cl->alist); in cl_is_active()
322 struct qfq_class *cl) in qfq_add_to_agg() argument
324 cl->agg = agg; in qfq_add_to_agg()
327 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg()
328 list_add_tail(&cl->alist, &agg->active); in qfq_add_to_agg()
330 cl && q->in_serv_agg != agg) /* agg was inactive */ in qfq_add_to_agg()
350 static void qfq_deactivate_class(struct qfq_sched *q, struct qfq_class *cl) in qfq_deactivate_class() argument
352 struct qfq_aggregate *agg = cl->agg; in qfq_deactivate_class()
355 list_del_init(&cl->alist); /* remove from RR queue of the aggregate */ in qfq_deactivate_class()
[all …]
H A Dsch_ets.c77 static bool cl_is_active(struct ets_class *cl) in cl_is_active() argument
79 return !list_empty(&cl->alist); in cl_is_active()
104 static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl) in ets_class_id() argument
107 int band = cl - q->classes; in ets_class_id()
201 static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl) in ets_class_is_strict() argument
203 unsigned int band = cl - q->classes; in ets_class_is_strict()
212 struct ets_class *cl = ets_class_from_arg(sch, *arg); in ets_class_change() local
222 if (!cl) { in ets_class_change()
240 if (ets_class_is_strict(q, cl)) { in ets_class_change()
251 cl->quantum = quantum; in ets_class_change()
[all …]
/linux/drivers/misc/mei/
H A Dclient.c319 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue()
331 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue()
332 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue()
345 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument
350 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp()
367 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument
379 cb->cl = cl; in mei_io_cb_init()
395 const struct mei_cl *cl) in mei_io_list_flush_cl() argument
400 if (cl == cb->cl) { in mei_io_list_flush_cl()
416 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument
[all …]
H A Dmain.c50 struct mei_cl *cl; in mei_open() local
68 cl = mei_cl_alloc_linked(dev); in mei_open()
69 if (IS_ERR(cl)) { in mei_open()
70 err = PTR_ERR(cl); in mei_open()
74 cl->fp = file; in mei_open()
75 file->private_data = cl; in mei_open()
94 static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl, in mei_cl_vtag_remove_by_fp() argument
99 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) { in mei_cl_vtag_remove_by_fp()
118 struct mei_cl *cl = file->private_data; in mei_release() local
122 if (WARN_ON(!cl || !cl->dev)) in mei_release()
[all …]
H A Dinterrupt.c32 struct mei_cl *cl; in mei_irq_compl_handler() local
35 cl = cb->cl; in mei_irq_compl_handler()
38 cl_dbg(dev, cl, "completing call back.\n"); in mei_irq_compl_handler()
39 mei_cl_complete(cl, cb); in mei_irq_compl_handler()
52 static inline int mei_cl_hbm_equal(struct mei_cl *cl, in mei_cl_hbm_equal() argument
55 return mei_cl_host_addr(cl) == mei_hdr->host_addr && in mei_cl_hbm_equal()
56 mei_cl_me_id(cl) == mei_hdr->me_addr; in mei_cl_hbm_equal()
93 static int mei_cl_irq_read_msg(struct mei_cl *cl, in mei_cl_irq_read_msg() argument
98 struct mei_device *dev = cl->dev; in mei_cl_irq_read_msg()
115 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg()
[all …]
H A Dbus.c35 ssize_t __mei_cl_send(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send() argument
38 return __mei_cl_send_timeout(cl, buf, length, vtag, mode, MAX_SCHEDULE_TIMEOUT); in __mei_cl_send()
55 ssize_t __mei_cl_send_timeout(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send_timeout() argument
62 if (WARN_ON(!cl || !cl->dev)) in __mei_cl_send_timeout()
65 bus = cl->dev; in __mei_cl_send_timeout()
74 if (!mei_cl_is_connected(cl)) { in __mei_cl_send_timeout()
80 if (!mei_me_cl_is_active(cl->me_cl)) { in __mei_cl_send_timeout()
87 rets = mei_cl_vt_support_check(cl); in __mei_cl_send_timeout()
92 if (length > mei_cl_mtu(cl)) { in __mei_cl_send_timeout()
97 while (cl->tx_cb_queued >= bus->tx_queue_limit) { in __mei_cl_send_timeout()
[all …]
H A Dhbm.c150 void mei_hbm_cl_hdr(struct mei_cl *cl, u8 hbm_cmd, void *buf, size_t len) in mei_hbm_cl_hdr() argument
157 cmd->host_addr = mei_cl_host_addr(cl); in mei_hbm_cl_hdr()
158 cmd->me_addr = mei_cl_me_id(cl); in mei_hbm_cl_hdr()
172 static inline int mei_hbm_cl_write(struct mei_device *dev, struct mei_cl *cl, in mei_hbm_cl_write() argument
178 mei_hbm_cl_hdr(cl, hbm_cmd, buf, len); in mei_hbm_cl_write()
193 bool mei_hbm_cl_addr_equal(struct mei_cl *cl, struct mei_hbm_cl_cmd *cmd) in mei_hbm_cl_addr_equal() argument
195 return mei_cl_host_addr(cl) == cmd->host_addr && in mei_hbm_cl_addr_equal()
196 mei_cl_me_id(cl) == cmd->me_addr; in mei_hbm_cl_addr_equal()
211 struct mei_cl *cl; in mei_hbm_cl_find_by_cmd() local
213 list_for_each_entry(cl, &dev->file_list, link) in mei_hbm_cl_find_by_cmd()
[all …]
/linux/sound/soc/intel/avs/
H A Dcldma.c60 void hda_cldma_fill(struct hda_cldma *cl) in hda_cldma_fill() argument
64 if (cl->remaining > cl->buffer_size) in hda_cldma_fill()
65 size = cl->buffer_size; in hda_cldma_fill()
67 size = cl->remaining; in hda_cldma_fill()
69 offset = snd_hdac_stream_readl(cl, CL_SD_SPIB); in hda_cldma_fill()
70 if (offset + size > cl->buffer_size) { in hda_cldma_fill()
73 ss = cl->buffer_size - offset; in hda_cldma_fill()
74 memcpy(cl->dmab_data.area + offset, cl->position, ss); in hda_cldma_fill()
77 cl->position += ss; in hda_cldma_fill()
78 cl->remaining -= ss; in hda_cldma_fill()
[all …]
H A Dcldma.h18 void hda_cldma_fill(struct hda_cldma *cl);
19 void hda_cldma_transfer(struct hda_cldma *cl, unsigned long start_delay);
21 int hda_cldma_start(struct hda_cldma *cl);
22 int hda_cldma_stop(struct hda_cldma *cl);
23 int hda_cldma_reset(struct hda_cldma *cl);
25 void hda_cldma_set_data(struct hda_cldma *cl, void *data, unsigned int size);
26 void hda_cldma_setup(struct hda_cldma *cl);
27 void hda_cldma_interrupt(struct hda_cldma *cl);
28 int hda_cldma_init(struct hda_cldma *cl, struct hdac_bus *bus, void __iomem *dsp_ba,
30 void hda_cldma_free(struct hda_cldma *cl);
/linux/include/linux/
H A Dclosure.h171 void closure_sub(struct closure *cl, int v);
172 void closure_put(struct closure *cl);
174 bool closure_wait(struct closure_waitlist *list, struct closure *cl);
175 void __closure_sync(struct closure *cl);
177 static inline unsigned closure_nr_remaining(struct closure *cl) in closure_nr_remaining() argument
179 return atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK; in closure_nr_remaining()
188 static inline void closure_sync(struct closure *cl) in closure_sync() argument
191 BUG_ON(closure_nr_remaining(cl) != 1 && !cl->closure_get_happened); in closure_sync()
194 if (cl->closure_get_happened) in closure_sync()
195 __closure_sync(cl); in closure_sync()
[all …]
/linux/lib/
H A Dclosure.c30 static inline void closure_put_after_sub(struct closure *cl, int flags) in closure_put_after_sub() argument
37 cl->closure_get_happened = false; in closure_put_after_sub()
39 if (cl->fn && !(flags & CLOSURE_DESTRUCTOR)) { in closure_put_after_sub()
40 atomic_set(&cl->remaining, in closure_put_after_sub()
42 closure_queue(cl); in closure_put_after_sub()
44 struct closure *parent = cl->parent; in closure_put_after_sub()
45 closure_fn *destructor = cl->fn; in closure_put_after_sub()
47 closure_debug_destroy(cl); in closure_put_after_sub()
50 destructor(&cl->work); in closure_put_after_sub()
59 void closure_sub(struct closure *cl, int v) in closure_sub() argument
[all …]
/linux/drivers/hsi/
H A Dhsi_core.c60 struct hsi_client *cl = to_hsi_client(dev); in hsi_client_release() local
62 kfree(cl->tx_cfg.channels); in hsi_client_release()
63 kfree(cl->rx_cfg.channels); in hsi_client_release()
64 kfree(cl); in hsi_client_release()
70 struct hsi_client *cl; in hsi_new_client() local
73 cl = kzalloc_obj(*cl); in hsi_new_client()
74 if (!cl) in hsi_new_client()
77 cl->tx_cfg = info->tx_cfg; in hsi_new_client()
78 if (cl->tx_cfg.channels) { in hsi_new_client()
79 size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels); in hsi_new_client()
[all …]
/linux/drivers/hsi/clients/
H A Dssi_protocol.c139 struct hsi_client *cl; member
217 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); in ssip_release_cmd()
219 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); in ssip_release_cmd()
331 if (slave->device.parent == ssi->cl->device.parent) { in ssip_slave_get_master()
332 master = ssi->cl; in ssip_slave_get_master()
385 static void ssip_reset(struct hsi_client *cl) in ssip_reset() argument
387 struct ssi_protocol *ssi = hsi_client_drvdata(cl); in ssip_reset()
393 hsi_flush(cl); in ssip_reset()
396 hsi_stop_tx(cl); in ssip_reset()
399 ssi_waketest(cl, 0); /* FIXME: To be removed */ in ssip_reset()
[all …]
H A Dhsi_char.c85 struct hsi_client *cl; member
105 struct hsi_client *cl; member
227 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_rx_completed()
247 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_tx_completed()
267 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_req_destructor()
275 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_received()
294 hsi_flush(msg->cl); in hsc_break_received()
295 ret = hsi_async_read(msg->cl, msg); in hsc_break_received()
300 static int hsc_break_request(struct hsi_client *cl) in hsc_break_request() argument
302 struct hsc_client_data *cl_data = hsi_client_drvdata(cl); in hsc_break_request()
[all …]
/linux/drivers/clk/
H A Dclkdev.c38 struct clk_lookup *p, *cl = NULL; in clk_find() local
62 cl = p; in clk_find()
69 return cl; in clk_find()
74 struct clk_lookup *cl; in clk_find_hw() local
78 cl = clk_find(dev_id, con_id); in clk_find_hw()
79 if (cl) in clk_find_hw()
80 hw = cl->clk_hw; in clk_find_hw()
121 static void __clkdev_add(struct clk_lookup *cl) in __clkdev_add() argument
124 list_add_tail(&cl->node, &clocks); in __clkdev_add()
128 void clkdev_add(struct clk_lookup *cl) in clkdev_add() argument
[all …]
/linux/include/linux/hsi/
H A Dhsi.h138 static inline void hsi_client_set_drvdata(struct hsi_client *cl, void *data) in hsi_client_set_drvdata() argument
140 dev_set_drvdata(&cl->device, data); in hsi_client_set_drvdata()
143 static inline void *hsi_client_drvdata(struct hsi_client *cl) in hsi_client_drvdata() argument
145 return dev_get_drvdata(&cl->device); in hsi_client_drvdata()
148 int hsi_register_port_event(struct hsi_client *cl,
150 int hsi_unregister_port_event(struct hsi_client *cl);
187 struct hsi_client *cl; member
230 int (*setup)(struct hsi_client *cl);
231 int (*flush)(struct hsi_client *cl);
232 int (*start_tx)(struct hsi_client *cl);
[all …]
/linux/drivers/md/bcache/
H A Drequest.c60 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK()
67 op->flush_journal ? cl : NULL); in CLOSURE_CALLBACK()
82 continue_at(cl, bch_data_insert_start, op->wq); in CLOSURE_CALLBACK()
87 closure_return(cl); in CLOSURE_CALLBACK()
108 static void bch_data_invalidate(struct closure *cl) in bch_data_invalidate() argument
110 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_invalidate()
136 continue_at(cl, bch_data_insert_keys, op->wq); in bch_data_invalidate()
141 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK()
166 bch_data_insert_keys(&cl->work); in CLOSURE_CALLBACK()
171 struct closure *cl = bio->bi_private; in bch_data_insert_endio() local
[all …]
/linux/include/linux/soc/samsung/
H A Dexynos-regs-pmu.h718 #define GS101_CLUSTER_CPU_OFFSET(cl, cpu) ((cl) + ((cpu) * 0x80)) argument
719 #define GS101_CLUSTER_CPU_CONFIGURATION(cl, cpu) \ argument
720 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x00)
721 #define GS101_CLUSTER_CPU_STATUS(cl, cpu) \ argument
722 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x04)
723 #define GS101_CLUSTER_CPU_STATES(cl, cpu) \ argument
724 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x08)
725 #define GS101_CLUSTER_CPU_OPTION(cl, cpu) \ argument
726 (GS101_CLUSTER_CPU_OFFSET(cl, cp
727 GS101_CLUSTER_CPU_OUT(cl,cpu) global() argument
729 GS101_CLUSTER_CPU_IN(cl,cpu) global() argument
731 GS101_CLUSTER_CPU_INT_IN(cl,cpu) global() argument
733 GS101_CLUSTER_CPU_INT_EN(cl,cpu) global() argument
735 GS101_CLUSTER_CPU_INT_TYPE(cl,cpu) global() argument
737 GS101_CLUSTER_CPU_INT_DIR(cl,cpu) global() argument
740 GS101_CLUSTER_NONCPU_OFFSET(cl) global() argument
741 GS101_CLUSTER_NONCPU_CONFIGURATION(cl) global() argument
743 GS101_CLUSTER_NONCPU_STATUS(cl) global() argument
745 GS101_CLUSTER_NONCPU_STATES(cl) global() argument
747 GS101_CLUSTER_NONCPU_OPTION(cl) global() argument
749 GS101_CLUSTER_NONCPU_OUT(cl) global() argument
751 GS101_CLUSTER_NONCPU_IN(cl) global() argument
753 GS101_CLUSTER_NONCPU_INT_IN(cl) global() argument
755 GS101_CLUSTER_NONCPU_INT_EN(cl) global() argument
757 GS101_CLUSTER_NONCPU_INT_TYPE(cl) global() argument
759 GS101_CLUSTER_NONCPU_INT_DIR(cl) global() argument
761 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_OUT(cl) global() argument
763 GS101_CLUSTER_NONCPU_DUALRAIL_POS_OUT(cl) global() argument
765 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_IN(cl) global() argument
[all...]
/linux/arch/x86/math-emu/
H A Dwm_shrx.S50 shrd %cl,%ebx,%eax
51 shrd %cl,%edx,%ebx
52 shr %cl,%edx
64 subb $32,%cl
67 shrd %cl,%edx,%eax
68 shr %cl,%edx
79 subb $64,%cl
81 shr %cl,%eax
131 subb $32,%cl
135 shrd %cl,%eax,%ebx
[all …]

12345678910>>...13