| /linux/drivers/hid/intel-ish-hid/ishtp/ |
| H A D | client.c | 23 static void ishtp_read_list_flush(struct ishtp_cl *cl) in ishtp_read_list_flush() argument 29 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush() 30 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush() 31 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush() 33 spin_lock(&cl->free_list_spinlock); in ishtp_read_list_flush() 34 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_read_list_flush() 35 spin_unlock(&cl->free_list_spinlock); in ishtp_read_list_flush() 37 spin_unlock_irqrestore(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush() 49 int ishtp_cl_flush_queues(struct ishtp_cl *cl) in ishtp_cl_flush_queues() argument 51 if (WARN_ON(!cl || !cl->dev)) in ishtp_cl_flush_queues() [all …]
|
| H A D | client-buffers.c | 19 int ishtp_cl_alloc_rx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_rx_ring() argument 21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring() 27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring() 28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring() 36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring() 38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring() 45 ishtp_cl_free_rx_ring(cl); in ishtp_cl_alloc_rx_ring() 57 int ishtp_cl_alloc_tx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_tx_ring() argument [all …]
|
| H A D | hbm.c | 55 static inline void ishtp_hbm_cl_hdr(struct ishtp_cl *cl, uint8_t hbm_cmd, in ishtp_hbm_cl_hdr() argument 63 cmd->host_addr = cl->host_client_id; in ishtp_hbm_cl_hdr() 64 cmd->fw_addr = cl->fw_client_id; in ishtp_hbm_cl_hdr() 76 static inline bool ishtp_hbm_cl_addr_equal(struct ishtp_cl *cl, void *buf) in ishtp_hbm_cl_addr_equal() argument 80 return cl->host_client_id == cmd->host_addr && in ishtp_hbm_cl_addr_equal() 81 cl->fw_client_id == cmd->fw_addr; in ishtp_hbm_cl_addr_equal() 262 struct ishtp_cl *cl) in ishtp_hbm_cl_flow_control_req() argument 270 spin_lock_irqsave(&cl->fc_spinlock, flags); in ishtp_hbm_cl_flow_control_req() 273 ishtp_hbm_cl_hdr(cl, ISHTP_FLOW_CONTROL_CMD, &flow_ctrl, len); in ishtp_hbm_cl_flow_control_req() 279 if (cl->out_flow_ctrl_creds) { in ishtp_hbm_cl_flow_control_req() [all …]
|
| /linux/net/sched/ |
| H A D | sch_hfsc.c | 178 static bool cl_in_el_or_vttree(struct hfsc_class *cl) in cl_in_el_or_vttree() argument 180 return ((cl->cl_flags & HFSC_FSC) && cl->cl_nactive) || in cl_in_el_or_vttree() 181 ((cl->cl_flags & HFSC_RSC) && !RB_EMPTY_NODE(&cl->el_node)); in cl_in_el_or_vttree() 190 eltree_insert(struct hfsc_class *cl) in eltree_insert() argument 192 struct rb_node **p = &cl->sched->eligible.rb_node; in eltree_insert() 199 if (cl->cl_e >= cl1->cl_e) in eltree_insert() 204 rb_link_node(&cl->el_node, parent, p); in eltree_insert() 205 rb_insert_color(&cl in eltree_insert() 209 eltree_remove(struct hfsc_class * cl) eltree_remove() argument 218 eltree_update(struct hfsc_class * cl) eltree_update() argument 228 struct hfsc_class *p, *cl = NULL; eltree_get_mindl() local 258 vttree_insert(struct hfsc_class * cl) vttree_insert() argument 277 vttree_remove(struct hfsc_class * cl) vttree_remove() argument 283 vttree_update(struct hfsc_class * cl) vttree_update() argument 290 vttree_firstfit(struct hfsc_class * cl,u64 cur_time) vttree_firstfit() argument 307 vttree_get_minvt(struct hfsc_class * cl,u64 cur_time) vttree_get_minvt() argument 327 cftree_insert(struct hfsc_class * cl) cftree_insert() argument 346 cftree_remove(struct hfsc_class * cl) cftree_remove() argument 352 cftree_update(struct hfsc_class * cl) cftree_update() argument 617 init_ed(struct hfsc_class * cl,unsigned int next_len) init_ed() argument 643 update_ed(struct hfsc_class * cl,unsigned int next_len) update_ed() argument 652 update_d(struct hfsc_class * cl,unsigned int next_len) update_d() argument 658 update_cfmin(struct hfsc_class * cl) update_cfmin() argument 672 init_vf(struct hfsc_class * cl,unsigned int len) init_vf() argument 751 update_vf(struct hfsc_class * cl,unsigned int len,u64 cur_time) update_vf() argument 839 hfsc_adjust_levels(struct hfsc_class * cl) hfsc_adjust_levels() argument 867 hfsc_change_rsc(struct hfsc_class * cl,struct tc_service_curve * rsc,u64 cur_time) hfsc_change_rsc() argument 881 hfsc_change_fsc(struct hfsc_class * cl,struct tc_service_curve * fsc) hfsc_change_fsc() argument 889 hfsc_change_usc(struct hfsc_class * cl,struct tc_service_curve * usc,u64 cur_time) hfsc_change_usc() argument 898 hfsc_upgrade_rt(struct hfsc_class * cl) hfsc_upgrade_rt() argument 917 struct hfsc_class *cl = (struct hfsc_class *)*arg; hfsc_change_class() local 1091 hfsc_destroy_class(struct Qdisc * sch,struct hfsc_class * cl) hfsc_destroy_class() argument 1107 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_delete_class() local 1133 struct hfsc_class *head, *cl; hfsc_classify() local 1188 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_graft_class() local 1206 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_class_leaf() local 1217 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_qlen_notify() local 1238 struct hfsc_class *cl = hfsc_find_class(classid, sch); hfsc_bind_tcf() local 1252 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_unbind_tcf() local 1261 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_tcf_block() local 1287 hfsc_dump_curves(struct sk_buff * skb,struct hfsc_class * cl) hfsc_dump_curves() argument 1311 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_dump_class() local 1336 struct hfsc_class *cl = (struct hfsc_class *)arg; hfsc_dump_class_stats() local 1360 struct hfsc_class *cl; hfsc_walk() local 1379 struct hfsc_class *cl; hfsc_schedule_watchdog() local 1453 hfsc_reset_class(struct hfsc_class * cl) hfsc_reset_class() argument 1486 struct hfsc_class *cl; hfsc_reset_qdisc() local 1502 struct hfsc_class *cl; hfsc_destroy_qdisc() local 1541 struct hfsc_class *cl; hfsc_enqueue() local 1588 struct hfsc_class *cl; hfsc_dequeue() local [all...] |
| H A D | sch_htb.c | 223 struct htb_class *cl; in htb_classify() local 234 cl = htb_find(skb->priority, sch); in htb_classify() 235 if (cl) { in htb_classify() 236 if (cl->level == 0) in htb_classify() 237 return cl; in htb_classify() 239 tcf = rcu_dereference_bh(cl->filter_list); in htb_classify() 257 cl = (void *)res.class; in htb_classify() 258 if (!cl) { in htb_classify() 261 cl = htb_find(res.classid, sch); in htb_classify() 262 if (!cl) in htb_classify() [all …]
|
| H A D | sch_drr.c | 38 static bool cl_is_active(struct drr_class *cl) in cl_is_active() argument 40 return !list_empty(&cl->alist); in cl_is_active() 63 struct drr_class *cl = (struct drr_class *)*arg; in drr_change_class() local 88 if (cl != NULL) { in drr_change_class() 90 err = gen_replace_estimator(&cl->bstats, NULL, in drr_change_class() 91 &cl->rate_est, in drr_change_class() 102 cl->quantum = quantum; in drr_change_class() 108 cl = kzalloc_obj(struct drr_class); in drr_change_class() 109 if (cl == NULL) in drr_change_class() 112 gnet_stats_basic_sync_init(&cl->bstats); in drr_change_class() [all …]
|
| H A D | sch_qfq.c | 205 static bool cl_is_active(struct qfq_class *cl) in cl_is_active() argument 207 return !list_empty(&cl->alist); in cl_is_active() 322 struct qfq_class *cl) in qfq_add_to_agg() argument 324 cl->agg = agg; in qfq_add_to_agg() 327 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg() 328 list_add_tail(&cl->alist, &agg->active); in qfq_add_to_agg() 330 cl && q->in_serv_agg != agg) /* agg was inactive */ in qfq_add_to_agg() 350 static void qfq_deactivate_class(struct qfq_sched *q, struct qfq_class *cl) in qfq_deactivate_class() argument 352 struct qfq_aggregate *agg = cl->agg; in qfq_deactivate_class() 355 list_del_init(&cl->alist); /* remove from RR queue of the aggregate */ in qfq_deactivate_class() [all …]
|
| H A D | sch_ets.c | 77 static bool cl_is_active(struct ets_class *cl) in cl_is_active() argument 79 return !list_empty(&cl->alist); in cl_is_active() 104 static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl) in ets_class_id() argument 107 int band = cl - q->classes; in ets_class_id() 201 static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl) in ets_class_is_strict() argument 203 unsigned int band = cl - q->classes; in ets_class_is_strict() 212 struct ets_class *cl = ets_class_from_arg(sch, *arg); in ets_class_change() local 222 if (!cl) { in ets_class_change() 240 if (ets_class_is_strict(q, cl)) { in ets_class_change() 251 cl->quantum = quantum; in ets_class_change() [all …]
|
| /linux/drivers/misc/mei/ |
| H A D | client.c | 319 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue() 331 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue() 332 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue() 345 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument 350 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp() 367 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument 379 cb->cl = cl; in mei_io_cb_init() 395 const struct mei_cl *cl) in mei_io_list_flush_cl() argument 400 if (cl == cb->cl) { in mei_io_list_flush_cl() 416 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument [all …]
|
| H A D | main.c | 50 struct mei_cl *cl; in mei_open() local 68 cl = mei_cl_alloc_linked(dev); in mei_open() 69 if (IS_ERR(cl)) { in mei_open() 70 err = PTR_ERR(cl); in mei_open() 74 cl->fp = file; in mei_open() 75 file->private_data = cl; in mei_open() 94 static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl, in mei_cl_vtag_remove_by_fp() argument 99 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) { in mei_cl_vtag_remove_by_fp() 118 struct mei_cl *cl = file->private_data; in mei_release() local 122 if (WARN_ON(!cl || !cl->dev)) in mei_release() [all …]
|
| H A D | interrupt.c | 32 struct mei_cl *cl; in mei_irq_compl_handler() local 35 cl = cb->cl; in mei_irq_compl_handler() 38 cl_dbg(dev, cl, "completing call back.\n"); in mei_irq_compl_handler() 39 mei_cl_complete(cl, cb); in mei_irq_compl_handler() 52 static inline int mei_cl_hbm_equal(struct mei_cl *cl, in mei_cl_hbm_equal() argument 55 return mei_cl_host_addr(cl) == mei_hdr->host_addr && in mei_cl_hbm_equal() 56 mei_cl_me_id(cl) == mei_hdr->me_addr; in mei_cl_hbm_equal() 93 static int mei_cl_irq_read_msg(struct mei_cl *cl, in mei_cl_irq_read_msg() argument 98 struct mei_device *dev = cl->dev; in mei_cl_irq_read_msg() 115 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg() [all …]
|
| H A D | bus.c | 35 ssize_t __mei_cl_send(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send() argument 38 return __mei_cl_send_timeout(cl, buf, length, vtag, mode, MAX_SCHEDULE_TIMEOUT); in __mei_cl_send() 55 ssize_t __mei_cl_send_timeout(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag, in __mei_cl_send_timeout() argument 62 if (WARN_ON(!cl || !cl->dev)) in __mei_cl_send_timeout() 65 bus = cl->dev; in __mei_cl_send_timeout() 74 if (!mei_cl_is_connected(cl)) { in __mei_cl_send_timeout() 80 if (!mei_me_cl_is_active(cl->me_cl)) { in __mei_cl_send_timeout() 87 rets = mei_cl_vt_support_check(cl); in __mei_cl_send_timeout() 92 if (length > mei_cl_mtu(cl)) { in __mei_cl_send_timeout() 97 while (cl->tx_cb_queued >= bus->tx_queue_limit) { in __mei_cl_send_timeout() [all …]
|
| H A D | hbm.c | 150 void mei_hbm_cl_hdr(struct mei_cl *cl, u8 hbm_cmd, void *buf, size_t len) in mei_hbm_cl_hdr() argument 157 cmd->host_addr = mei_cl_host_addr(cl); in mei_hbm_cl_hdr() 158 cmd->me_addr = mei_cl_me_id(cl); in mei_hbm_cl_hdr() 172 static inline int mei_hbm_cl_write(struct mei_device *dev, struct mei_cl *cl, in mei_hbm_cl_write() argument 178 mei_hbm_cl_hdr(cl, hbm_cmd, buf, len); in mei_hbm_cl_write() 193 bool mei_hbm_cl_addr_equal(struct mei_cl *cl, struct mei_hbm_cl_cmd *cmd) in mei_hbm_cl_addr_equal() argument 195 return mei_cl_host_addr(cl) == cmd->host_addr && in mei_hbm_cl_addr_equal() 196 mei_cl_me_id(cl) == cmd->me_addr; in mei_hbm_cl_addr_equal() 211 struct mei_cl *cl; in mei_hbm_cl_find_by_cmd() local 213 list_for_each_entry(cl, &dev->file_list, link) in mei_hbm_cl_find_by_cmd() [all …]
|
| /linux/sound/soc/intel/avs/ |
| H A D | cldma.c | 60 void hda_cldma_fill(struct hda_cldma *cl) in hda_cldma_fill() argument 64 if (cl->remaining > cl->buffer_size) in hda_cldma_fill() 65 size = cl->buffer_size; in hda_cldma_fill() 67 size = cl->remaining; in hda_cldma_fill() 69 offset = snd_hdac_stream_readl(cl, CL_SD_SPIB); in hda_cldma_fill() 70 if (offset + size > cl->buffer_size) { in hda_cldma_fill() 73 ss = cl->buffer_size - offset; in hda_cldma_fill() 74 memcpy(cl->dmab_data.area + offset, cl->position, ss); in hda_cldma_fill() 77 cl->position += ss; in hda_cldma_fill() 78 cl->remaining -= ss; in hda_cldma_fill() [all …]
|
| H A D | cldma.h | 18 void hda_cldma_fill(struct hda_cldma *cl); 19 void hda_cldma_transfer(struct hda_cldma *cl, unsigned long start_delay); 21 int hda_cldma_start(struct hda_cldma *cl); 22 int hda_cldma_stop(struct hda_cldma *cl); 23 int hda_cldma_reset(struct hda_cldma *cl); 25 void hda_cldma_set_data(struct hda_cldma *cl, void *data, unsigned int size); 26 void hda_cldma_setup(struct hda_cldma *cl); 27 void hda_cldma_interrupt(struct hda_cldma *cl); 28 int hda_cldma_init(struct hda_cldma *cl, struct hdac_bus *bus, void __iomem *dsp_ba, 30 void hda_cldma_free(struct hda_cldma *cl);
|
| /linux/include/linux/ |
| H A D | closure.h | 171 void closure_sub(struct closure *cl, int v); 172 void closure_put(struct closure *cl); 174 bool closure_wait(struct closure_waitlist *list, struct closure *cl); 175 void __closure_sync(struct closure *cl); 177 static inline unsigned closure_nr_remaining(struct closure *cl) in closure_nr_remaining() argument 179 return atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK; in closure_nr_remaining() 188 static inline void closure_sync(struct closure *cl) in closure_sync() argument 191 BUG_ON(closure_nr_remaining(cl) != 1 && !cl->closure_get_happened); in closure_sync() 194 if (cl->closure_get_happened) in closure_sync() 195 __closure_sync(cl); in closure_sync() [all …]
|
| /linux/lib/ |
| H A D | closure.c | 30 static inline void closure_put_after_sub(struct closure *cl, int flags) in closure_put_after_sub() argument 37 cl->closure_get_happened = false; in closure_put_after_sub() 39 if (cl->fn && !(flags & CLOSURE_DESTRUCTOR)) { in closure_put_after_sub() 40 atomic_set(&cl->remaining, in closure_put_after_sub() 42 closure_queue(cl); in closure_put_after_sub() 44 struct closure *parent = cl->parent; in closure_put_after_sub() 45 closure_fn *destructor = cl->fn; in closure_put_after_sub() 47 closure_debug_destroy(cl); in closure_put_after_sub() 50 destructor(&cl->work); in closure_put_after_sub() 59 void closure_sub(struct closure *cl, int v) in closure_sub() argument [all …]
|
| /linux/drivers/hsi/ |
| H A D | hsi_core.c | 60 struct hsi_client *cl = to_hsi_client(dev); in hsi_client_release() local 62 kfree(cl->tx_cfg.channels); in hsi_client_release() 63 kfree(cl->rx_cfg.channels); in hsi_client_release() 64 kfree(cl); in hsi_client_release() 70 struct hsi_client *cl; in hsi_new_client() local 73 cl = kzalloc_obj(*cl); in hsi_new_client() 74 if (!cl) in hsi_new_client() 77 cl->tx_cfg = info->tx_cfg; in hsi_new_client() 78 if (cl->tx_cfg.channels) { in hsi_new_client() 79 size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels); in hsi_new_client() [all …]
|
| /linux/drivers/hsi/clients/ |
| H A D | ssi_protocol.c | 139 struct hsi_client *cl; member 217 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); in ssip_release_cmd() 219 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); in ssip_release_cmd() 331 if (slave->device.parent == ssi->cl->device.parent) { in ssip_slave_get_master() 332 master = ssi->cl; in ssip_slave_get_master() 385 static void ssip_reset(struct hsi_client *cl) in ssip_reset() argument 387 struct ssi_protocol *ssi = hsi_client_drvdata(cl); in ssip_reset() 393 hsi_flush(cl); in ssip_reset() 396 hsi_stop_tx(cl); in ssip_reset() 399 ssi_waketest(cl, 0); /* FIXME: To be removed */ in ssip_reset() [all …]
|
| H A D | hsi_char.c | 85 struct hsi_client *cl; member 105 struct hsi_client *cl; member 227 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_rx_completed() 247 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_tx_completed() 267 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_req_destructor() 275 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_received() 294 hsi_flush(msg->cl); in hsc_break_received() 295 ret = hsi_async_read(msg->cl, msg); in hsc_break_received() 300 static int hsc_break_request(struct hsi_client *cl) in hsc_break_request() argument 302 struct hsc_client_data *cl_data = hsi_client_drvdata(cl); in hsc_break_request() [all …]
|
| /linux/drivers/clk/ |
| H A D | clkdev.c | 38 struct clk_lookup *p, *cl = NULL; in clk_find() local 62 cl = p; in clk_find() 69 return cl; in clk_find() 74 struct clk_lookup *cl; in clk_find_hw() local 78 cl = clk_find(dev_id, con_id); in clk_find_hw() 79 if (cl) in clk_find_hw() 80 hw = cl->clk_hw; in clk_find_hw() 121 static void __clkdev_add(struct clk_lookup *cl) in __clkdev_add() argument 124 list_add_tail(&cl->node, &clocks); in __clkdev_add() 128 void clkdev_add(struct clk_lookup *cl) in clkdev_add() argument [all …]
|
| /linux/include/linux/hsi/ |
| H A D | hsi.h | 138 static inline void hsi_client_set_drvdata(struct hsi_client *cl, void *data) in hsi_client_set_drvdata() argument 140 dev_set_drvdata(&cl->device, data); in hsi_client_set_drvdata() 143 static inline void *hsi_client_drvdata(struct hsi_client *cl) in hsi_client_drvdata() argument 145 return dev_get_drvdata(&cl->device); in hsi_client_drvdata() 148 int hsi_register_port_event(struct hsi_client *cl, 150 int hsi_unregister_port_event(struct hsi_client *cl); 187 struct hsi_client *cl; member 230 int (*setup)(struct hsi_client *cl); 231 int (*flush)(struct hsi_client *cl); 232 int (*start_tx)(struct hsi_client *cl); [all …]
|
| /linux/drivers/md/bcache/ |
| H A D | request.c | 60 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK() 67 op->flush_journal ? cl : NULL); in CLOSURE_CALLBACK() 82 continue_at(cl, bch_data_insert_start, op->wq); in CLOSURE_CALLBACK() 87 closure_return(cl); in CLOSURE_CALLBACK() 108 static void bch_data_invalidate(struct closure *cl) in bch_data_invalidate() argument 110 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_invalidate() 136 continue_at(cl, bch_data_insert_keys, op->wq); in bch_data_invalidate() 141 closure_type(op, struct data_insert_op, cl); in CLOSURE_CALLBACK() 166 bch_data_insert_keys(&cl->work); in CLOSURE_CALLBACK() 171 struct closure *cl = bio->bi_private; in bch_data_insert_endio() local [all …]
|
| /linux/include/linux/soc/samsung/ |
| H A D | exynos-regs-pmu.h | 718 #define GS101_CLUSTER_CPU_OFFSET(cl, cpu) ((cl) + ((cpu) * 0x80)) argument 719 #define GS101_CLUSTER_CPU_CONFIGURATION(cl, cpu) \ argument 720 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x00) 721 #define GS101_CLUSTER_CPU_STATUS(cl, cpu) \ argument 722 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x04) 723 #define GS101_CLUSTER_CPU_STATES(cl, cpu) \ argument 724 (GS101_CLUSTER_CPU_OFFSET(cl, cpu) + 0x08) 725 #define GS101_CLUSTER_CPU_OPTION(cl, cpu) \ argument 726 (GS101_CLUSTER_CPU_OFFSET(cl, cp 727 GS101_CLUSTER_CPU_OUT(cl,cpu) global() argument 729 GS101_CLUSTER_CPU_IN(cl,cpu) global() argument 731 GS101_CLUSTER_CPU_INT_IN(cl,cpu) global() argument 733 GS101_CLUSTER_CPU_INT_EN(cl,cpu) global() argument 735 GS101_CLUSTER_CPU_INT_TYPE(cl,cpu) global() argument 737 GS101_CLUSTER_CPU_INT_DIR(cl,cpu) global() argument 740 GS101_CLUSTER_NONCPU_OFFSET(cl) global() argument 741 GS101_CLUSTER_NONCPU_CONFIGURATION(cl) global() argument 743 GS101_CLUSTER_NONCPU_STATUS(cl) global() argument 745 GS101_CLUSTER_NONCPU_STATES(cl) global() argument 747 GS101_CLUSTER_NONCPU_OPTION(cl) global() argument 749 GS101_CLUSTER_NONCPU_OUT(cl) global() argument 751 GS101_CLUSTER_NONCPU_IN(cl) global() argument 753 GS101_CLUSTER_NONCPU_INT_IN(cl) global() argument 755 GS101_CLUSTER_NONCPU_INT_EN(cl) global() argument 757 GS101_CLUSTER_NONCPU_INT_TYPE(cl) global() argument 759 GS101_CLUSTER_NONCPU_INT_DIR(cl) global() argument 761 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_OUT(cl) global() argument 763 GS101_CLUSTER_NONCPU_DUALRAIL_POS_OUT(cl) global() argument 765 GS101_CLUSTER_NONCPU_DUALRAIL_CTRL_IN(cl) global() argument [all...] |
| /linux/arch/x86/math-emu/ |
| H A D | wm_shrx.S | 50 shrd %cl,%ebx,%eax 51 shrd %cl,%edx,%ebx 52 shr %cl,%edx 64 subb $32,%cl 67 shrd %cl,%edx,%eax 68 shr %cl,%edx 79 subb $64,%cl 81 shr %cl,%eax 131 subb $32,%cl 135 shrd %cl,%eax,%ebx [all …]
|