Home
last modified time | relevance | path

Searched full:cp (Results 1 – 25 of 802) sorted by relevance

12345678910>>...33

/linux/drivers/media/usb/pvrusb2/
H A Dpvrusb2-ioread.c39 static int pvr2_ioread_init(struct pvr2_ioread *cp) in pvr2_ioread_init() argument
43 cp->stream = NULL; in pvr2_ioread_init()
44 mutex_init(&cp->mutex); in pvr2_ioread_init()
47 cp->buffer_storage[idx] = kmalloc(BUFFER_SIZE,GFP_KERNEL); in pvr2_ioread_init()
48 if (!(cp->buffer_storage[idx])) break; in pvr2_ioread_init()
54 if (!(cp->buffer_storage[idx])) continue; in pvr2_ioread_init()
55 kfree(cp->buffer_storage[idx]); in pvr2_ioread_init()
62 static void pvr2_ioread_done(struct pvr2_ioread *cp) in pvr2_ioread_done() argument
66 pvr2_ioread_setup(cp,NULL); in pvr2_ioread_done()
68 if (!(cp->buffer_storage[idx])) continue; in pvr2_ioread_done()
[all …]
H A Dpvrusb2-context.c237 struct pvr2_channel *cp; in pvr2_context_reset_input_limits() local
241 for (cp = mp->mc_first; cp; cp = cp->mc_next) { in pvr2_context_reset_input_limits()
242 if (!cp->input_mask) continue; in pvr2_context_reset_input_limits()
243 tmsk &= cp->input_mask; in pvr2_context_reset_input_limits()
276 void pvr2_channel_init(struct pvr2_channel *cp,struct pvr2_context *mp) in pvr2_channel_init() argument
279 cp->hdw = mp->hdw; in pvr2_channel_init()
280 cp->mc_head = mp; in pvr2_channel_init()
281 cp->mc_next = NULL; in pvr2_channel_init()
282 cp->mc_prev = mp->mc_last; in pvr2_channel_init()
284 mp->mc_last->mc_next = cp; in pvr2_channel_init()
[all …]
/linux/drivers/net/ethernet/sun/
H A Dcassini.c111 * also, we need to make cp->lock finer-grained.
161 #define CAS_MAX_MTU min(((cp->page_size << 1) - 0x50), 9000)
230 static void cas_set_link_modes(struct cas *cp);
232 static inline void cas_lock_tx(struct cas *cp) in cas_lock_tx() argument
237 spin_lock_nested(&cp->tx_lock[i], i); in cas_lock_tx()
248 #define cas_lock_all_save(cp, flags) \ argument
250 struct cas *xxxcp = (cp); \
255 static inline void cas_unlock_tx(struct cas *cp) in cas_unlock_tx() argument
260 spin_unlock(&cp->tx_lock[i - 1]); in cas_unlock_tx()
263 #define cas_unlock_all_restore(cp, flags) \ argument
[all …]
/linux/net/netfilter/ipvs/
H A Dip_vs_conn.c143 static unsigned int ip_vs_conn_hashkey_conn(const struct ip_vs_conn *cp) in ip_vs_conn_hashkey_conn() argument
147 ip_vs_conn_fill_param(cp->ipvs, cp->af, cp->protocol, in ip_vs_conn_hashkey_conn()
148 &cp->caddr, cp->cport, NULL, 0, &p); in ip_vs_conn_hashkey_conn()
150 if (cp->pe) { in ip_vs_conn_hashkey_conn()
151 p.pe = cp->pe; in ip_vs_conn_hashkey_conn()
152 p.pe_data = cp->pe_data; in ip_vs_conn_hashkey_conn()
153 p.pe_data_len = cp in ip_vs_conn_hashkey_conn()
163 ip_vs_conn_hash(struct ip_vs_conn * cp) ip_vs_conn_hash() argument
199 ip_vs_conn_unhash(struct ip_vs_conn * cp) ip_vs_conn_unhash() argument
227 ip_vs_conn_unlink(struct ip_vs_conn * cp) ip_vs_conn_unlink() argument
266 struct ip_vs_conn *cp; __ip_vs_conn_in_get() local
295 struct ip_vs_conn *cp; ip_vs_conn_in_get() local
352 struct ip_vs_conn *cp; ip_vs_ct_in_get() local
404 struct ip_vs_conn *cp, *ret=NULL; ip_vs_conn_out_get() local
468 __ip_vs_conn_put_timer(struct ip_vs_conn * cp) __ip_vs_conn_put_timer() argument
477 ip_vs_conn_put(struct ip_vs_conn * cp) ip_vs_conn_put() argument
491 ip_vs_conn_fill_cport(struct ip_vs_conn * cp,__be16 cport) ip_vs_conn_fill_cport() argument
512 ip_vs_bind_xmit(struct ip_vs_conn * cp) ip_vs_bind_xmit() argument
543 ip_vs_bind_xmit_v6(struct ip_vs_conn * cp) ip_vs_bind_xmit_v6() argument
584 ip_vs_bind_dest(struct ip_vs_conn * cp,struct ip_vs_dest * dest) ip_vs_bind_dest() argument
651 ip_vs_try_bind_dest(struct ip_vs_conn * cp) ip_vs_try_bind_dest() argument
704 ip_vs_unbind_dest(struct ip_vs_conn * cp) ip_vs_unbind_dest() argument
814 struct ip_vs_conn *cp = container_of(head, struct ip_vs_conn, ip_vs_conn_rcu_free() local
823 ip_vs_conn_del(struct ip_vs_conn * cp) ip_vs_conn_del() argument
834 ip_vs_conn_del_put(struct ip_vs_conn * cp) ip_vs_conn_del_put() argument
849 struct ip_vs_conn *cp = timer_container_of(cp, t, timer); ip_vs_conn_expire() local
926 ip_vs_conn_expire_now(struct ip_vs_conn * cp) ip_vs_conn_expire_now() argument
945 struct ip_vs_conn *cp; ip_vs_conn_new() local
1056 struct ip_vs_conn *cp; ip_vs_conn_array() local
1098 struct ip_vs_conn *cp = v; ip_vs_conn_seq_next() local
1132 const struct ip_vs_conn *cp = v; ip_vs_conn_seq_show() local
1210 const struct ip_vs_conn *cp = v; ip_vs_conn_sync_seq_show() local
1268 todrop_entry(struct ip_vs_conn * cp) todrop_entry() argument
1296 ip_vs_conn_ops_mode(struct ip_vs_conn * cp) ip_vs_conn_ops_mode() argument
1310 struct ip_vs_conn *cp; ip_vs_random_dropentry() local
1380 struct ip_vs_conn *cp, *cp_c; ip_vs_conn_flush() local
1415 struct ip_vs_conn *cp, *cp_c; ip_vs_expire_nodest_conn_flush() local
[all...]
H A Dip_vs_proto_udp.c137 struct ip_vs_conn *cp, struct ip_vs_iphdr *iph) in udp_snat_handler() argument
145 if (cp->af == AF_INET6 && iph->fragoffs) in udp_snat_handler()
154 if (unlikely(cp->app != NULL)) { in udp_snat_handler()
158 if (!udp_csum_check(cp->af, skb, pp)) in udp_snat_handler()
164 if (!(ret = ip_vs_app_pkt_out(cp, skb, iph))) in udp_snat_handler()
174 udph->source = cp->vport; in udp_snat_handler()
180 udp_partial_csum_update(cp->af, udph, &cp->daddr, &cp->vaddr, in udp_snat_handler()
185 udp_fast_csum_update(cp->af, udph, &cp->daddr, &cp->vaddr, in udp_snat_handler()
186 cp->dport, cp->vport); in udp_snat_handler()
188 skb->ip_summed = cp->app ? in udp_snat_handler()
[all …]
H A Dip_vs_proto_tcp.c148 struct ip_vs_conn *cp, struct ip_vs_iphdr *iph) in tcp_snat_handler() argument
156 if (cp->af == AF_INET6 && iph->fragoffs) in tcp_snat_handler()
165 if (unlikely(cp->app != NULL)) { in tcp_snat_handler()
169 if (!tcp_csum_check(cp->af, skb, pp)) in tcp_snat_handler()
173 if (!(ret = ip_vs_app_pkt_out(cp, skb, iph))) in tcp_snat_handler()
183 tcph->source = cp->vport; in tcp_snat_handler()
187 tcp_partial_csum_update(cp->af, tcph, &cp->daddr, &cp->vaddr, in tcp_snat_handler()
192 tcp_fast_csum_update(cp->af, tcph, &cp->daddr, &cp->vaddr, in tcp_snat_handler()
193 cp->dport, cp->vport); in tcp_snat_handler()
195 skb->ip_summed = cp->app ? in tcp_snat_handler()
[all …]
H A Dip_vs_nfct.c71 ip_vs_update_conntrack(struct sk_buff *skb, struct ip_vs_conn *cp, int outin) in ip_vs_update_conntrack() argument
82 if (IP_VS_FWD_METHOD(cp) != IP_VS_CONN_F_MASQ) in ip_vs_update_conntrack()
86 if (cp->flags & IP_VS_CONN_F_ONE_PACKET) in ip_vs_update_conntrack()
94 if (cp->app && nf_ct_protonum(ct) == IPPROTO_TCP && in ip_vs_update_conntrack()
109 new_tuple.src.u3 = cp->daddr; in ip_vs_update_conntrack()
112 new_tuple.src.u.tcp.port = cp->dport; in ip_vs_update_conntrack()
114 new_tuple.dst.u3 = cp->vaddr; in ip_vs_update_conntrack()
117 new_tuple.dst.u.tcp.port = cp->vport; in ip_vs_update_conntrack()
128 IP_VS_DBG_BUF(7, "%s: Updated conntrack ct=%p for cp=" FMT_CONN "\n", in ip_vs_update_conntrack()
129 __func__, ct, ARG_CONN(cp)); in ip_vs_update_conntrack()
[all …]
H A Dip_vs_xmit.c591 struct ip_vs_conn *cp) in ip_vs_tunnel_xmit_prepare() argument
596 if (unlikely(cp->flags & IP_VS_CONN_F_NFCT)) in ip_vs_tunnel_xmit_prepare()
624 struct ip_vs_conn *cp, int local) in ip_vs_nat_send_or_cont() argument
629 if (likely(!(cp->flags & IP_VS_CONN_F_NFCT))) in ip_vs_nat_send_or_cont()
632 ip_vs_update_conntrack(skb, cp, 1); in ip_vs_nat_send_or_cont()
637 if (!local || cp->vport != cp->dport || in ip_vs_nat_send_or_cont()
638 !ip_vs_addr_equal(cp->af, &cp->vaddr, &cp->daddr)) in ip_vs_nat_send_or_cont()
645 NF_HOOK(pf, NF_INET_LOCAL_OUT, cp->ipvs->net, NULL, skb, in ip_vs_nat_send_or_cont()
655 struct ip_vs_conn *cp, int local) in ip_vs_send_or_cont() argument
660 if (likely(!(cp->flags & IP_VS_CONN_F_NFCT))) in ip_vs_send_or_cont()
[all …]
/linux/kernel/sched/
H A Dcpudeadline.c26 static void cpudl_heapify_down(struct cpudl *cp, int idx) in cpudl_heapify_down() argument
30 int orig_cpu = cp->elements[idx].cpu; in cpudl_heapify_down()
31 u64 orig_dl = cp->elements[idx].dl; in cpudl_heapify_down()
33 if (left_child(idx) >= cp->size) in cpudl_heapify_down()
45 if ((l < cp->size) && dl_time_before(orig_dl, in cpudl_heapify_down()
46 cp->elements[l].dl)) { in cpudl_heapify_down()
48 largest_dl = cp->elements[l].dl; in cpudl_heapify_down()
50 if ((r < cp->size) && dl_time_before(largest_dl, in cpudl_heapify_down()
51 cp->elements[r].dl)) in cpudl_heapify_down()
58 cp->elements[idx].cpu = cp->elements[largest].cpu; in cpudl_heapify_down()
[all …]
/linux/drivers/net/ethernet/realtek/
H A D8139cp.c1 /* 8139cp.c: A Linux PCI Ethernet driver for the RealTek 8139C+ chips. */
51 #define DRV_NAME "8139cp"
92 MODULE_PARM_DESC (debug, "8139cp: bitmapped message enable number");
98 MODULE_PARM_DESC (multicast_filter_limit, "8139cp: maximum number of filtered multicast addresses");
115 #define TX_BUFFS_AVAIL(CP) \ argument
116 (((CP)->tx_tail <= (CP)->tx_head) ? \
117 (CP)->tx_tail + (CP_TX_RING_SIZE - 1) - (CP)->tx_head : \
118 (CP)->tx_tail - (CP)->tx_head - 1)
355 #define cpr8(reg) readb(cp->regs + (reg))
356 #define cpr16(reg) readw(cp->regs + (reg))
[all …]
/linux/drivers/net/ethernet/broadcom/
H A Dcnic.c191 struct cnic_local *cp = dev->cnic_priv; in cnic_ctx_wr() local
192 struct cnic_eth_dev *ethdev = cp->ethdev; in cnic_ctx_wr()
206 struct cnic_local *cp = dev->cnic_priv; in cnic_ctx_tbl_wr() local
207 struct cnic_eth_dev *ethdev = cp->ethdev; in cnic_ctx_tbl_wr()
220 struct cnic_local *cp = dev->cnic_priv; in cnic_ring_ctl() local
221 struct cnic_eth_dev *ethdev = cp->ethdev; in cnic_ring_ctl()
238 struct cnic_local *cp = dev->cnic_priv; in cnic_reg_wr_ind() local
239 struct cnic_eth_dev *ethdev = cp->ethdev; in cnic_reg_wr_ind()
252 struct cnic_local *cp = dev->cnic_priv; in cnic_reg_rd_ind() local
253 struct cnic_eth_dev *ethdev = cp->ethdev; in cnic_reg_rd_ind()
[all …]
/linux/net/rds/
H A Dthreads.c74 void rds_connect_path_complete(struct rds_conn_path *cp, int curr) in rds_connect_path_complete() argument
76 if (!rds_conn_path_transition(cp, curr, RDS_CONN_UP)) { in rds_connect_path_complete()
80 atomic_read(&cp->cp_state)); in rds_connect_path_complete()
81 rds_conn_path_drop(cp, false); in rds_connect_path_complete()
86 cp->cp_conn, &cp->cp_conn->c_laddr, &cp->cp_conn->c_faddr); in rds_connect_path_complete()
88 cp->cp_reconnect_jiffies = 0; in rds_connect_path_complete()
89 set_bit(0, &cp->cp_conn->c_map_queued); in rds_connect_path_complete()
91 if (!rds_destroy_pending(cp->cp_conn)) { in rds_connect_path_complete()
92 queue_delayed_work(rds_wq, &cp->cp_send_w, 0); in rds_connect_path_complete()
93 queue_delayed_work(rds_wq, &cp->cp_recv_w, 0); in rds_connect_path_complete()
[all …]
H A Dconnection.c114 static void rds_conn_path_reset(struct rds_conn_path *cp) in rds_conn_path_reset()
116 struct rds_connection *conn = cp->cp_conn; in rds_conn_path_reset()
122 rds_send_path_reset(cp); in rds_conn_path_reset()
123 cp->cp_flags = 0; in rds_conn_path_reset()
132 struct rds_conn_path *cp, bool is_outgoing) in __rds_conn_path_init()
134 spin_lock_init(&cp->cp_lock); in __rds_conn_path_init()
135 cp->cp_next_tx_seq = 1; in __rds_conn_path_init()
136 init_waitqueue_head(&cp->cp_waitq); in __rds_conn_path_init()
137 INIT_LIST_HEAD(&cp->cp_send_queue); in __rds_conn_path_init()
138 INIT_LIST_HEAD(&cp in __rds_conn_path_init()
113 rds_conn_path_reset(struct rds_conn_path * cp) rds_conn_path_reset() argument
131 __rds_conn_path_init(struct rds_connection * conn,struct rds_conn_path * cp,bool is_outgoing) __rds_conn_path_init() argument
317 struct rds_conn_path *cp; __rds_conn_create() local
367 rds_conn_shutdown(struct rds_conn_path * cp) rds_conn_shutdown() argument
441 rds_conn_path_destroy(struct rds_conn_path * cp) rds_conn_path_destroy() argument
485 struct rds_conn_path *cp; rds_conn_destroy() local
558 struct rds_conn_path *cp; rds_conn_message_info_cmn() local
709 struct rds_conn_path *cp; rds_walk_conn_path_info() local
739 rds_conn_info_visitor(struct rds_conn_path * cp,void * buffer) rds_conn_info_visitor() argument
769 rds6_conn_info_visitor(struct rds_conn_path * cp,void * buffer) rds6_conn_info_visitor() argument
879 rds_conn_path_drop(struct rds_conn_path * cp,bool destroy) rds_conn_path_drop() argument
904 rds_conn_path_connect_if_down(struct rds_conn_path * cp) rds_conn_path_connect_if_down() argument
937 __rds_conn_path_error(struct rds_conn_path * cp,const char * fmt,...) __rds_conn_path_error() argument
[all...]
H A Dsend.c65 void rds_send_path_reset(struct rds_conn_path *cp) in rds_send_path_reset() argument
70 if (cp->cp_xmit_rm) { in rds_send_path_reset()
71 rm = cp->cp_xmit_rm; in rds_send_path_reset()
72 cp->cp_xmit_rm = NULL; in rds_send_path_reset()
81 cp->cp_xmit_sg = 0; in rds_send_path_reset()
82 cp->cp_xmit_hdr_off = 0; in rds_send_path_reset()
83 cp->cp_xmit_data_off = 0; in rds_send_path_reset()
84 cp->cp_xmit_atomic_sent = 0; in rds_send_path_reset()
85 cp->cp_xmit_rdma_sent = 0; in rds_send_path_reset()
86 cp in rds_send_path_reset()
104 acquire_in_xmit(struct rds_conn_path * cp) acquire_in_xmit() argument
109 release_in_xmit(struct rds_conn_path * cp) release_in_xmit() argument
136 rds_send_xmit(struct rds_conn_path * cp) rds_send_xmit() argument
685 rds_send_path_drop_acked(struct rds_conn_path * cp,u64 ack,is_acked_func is_acked) rds_send_path_drop_acked() argument
725 struct rds_conn_path *cp; rds_send_drop_to() local
818 rds_send_queue_rm(struct rds_sock * rs,struct rds_connection * conn,struct rds_conn_path * cp,struct rds_message * rm,__be16 sport,__be16 dport,int * queued) rds_send_queue_rm() argument
1421 rds_send_probe(struct rds_conn_path * cp,__be16 sport,__be16 dport,u8 h_flags) rds_send_probe() argument
1489 rds_send_pong(struct rds_conn_path * cp,__be16 dport) rds_send_pong() argument
1498 struct rds_conn_path *cp = &conn->c_path[cp_index]; rds_send_ping() local
[all...]
/linux/tools/perf/util/
H A Dcall-path.c14 static void call_path__init(struct call_path *cp, struct call_path *parent, in call_path__init() argument
17 cp->parent = parent; in call_path__init()
18 cp->sym = sym; in call_path__init()
19 cp->ip = sym ? 0 : ip; in call_path__init()
20 cp->db_id = 0; in call_path__init()
21 cp->in_kernel = in_kernel; in call_path__init()
22 RB_CLEAR_NODE(&cp->rb_node); in call_path__init()
23 cp->children = RB_ROOT; in call_path__init()
55 struct call_path *cp; in call_path__new() local
70 cp = &cpb->cp[n]; in call_path__new()
[all …]
/linux/scripts/
H A Dunifdef.c642 const char *cp; in parseline() local
659 cp = skipcomment(tline); in parseline()
661 if (*cp == '#') { in parseline()
664 cp = skipcomment(cp + 1); in parseline()
665 } else if (*cp != '\0') in parseline()
669 keyword = tline + (cp - tline); in parseline()
670 cp = skipsym(cp); in parseline()
671 kwlen = cp - keyword; in parseline()
673 if (strncmp(cp, "\\\r\n", 3) == 0 || in parseline()
674 strncmp(cp, "\\\n", 2) == 0) in parseline()
[all …]
/linux/arch/riscv/kvm/
H A Dvcpu_sbi_replace.c20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler() local
23 if (cp->a6 != SBI_EXT_TIME_SET_TIMER) { in kvm_sbi_ext_time_handler()
30 next_cycle = ((u64)cp->a1 << 32) | (u64)cp->a0; in kvm_sbi_ext_time_handler()
32 next_cycle = (u64)cp->a0; in kvm_sbi_ext_time_handler()
51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler() local
52 unsigned long hmask = cp->a0; in kvm_sbi_ext_ipi_handler()
53 unsigned long hbase = cp->a1; in kvm_sbi_ext_ipi_handler()
56 if (cp->a6 != SBI_EXT_IPI_SEND_IPI) { in kvm_sbi_ext_ipi_handler()
95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler() local
96 unsigned long hmask = cp->a0; in kvm_sbi_ext_rfence_handler()
[all …]
H A Dvcpu_sbi_v01.c24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler() local
28 switch (cp->a7) { in kvm_sbi_ext_v01_handler()
40 next_cycle = ((u64)cp->a1 << 32) | (u64)cp->a0; in kvm_sbi_ext_v01_handler()
42 next_cycle = (u64)cp->a0; in kvm_sbi_ext_v01_handler()
50 if (cp->a0) in kvm_sbi_ext_v01_handler()
51 hmask = kvm_riscv_vcpu_unpriv_read(vcpu, false, cp->a0, utrap); in kvm_sbi_ext_v01_handler()
72 if (cp->a0) in kvm_sbi_ext_v01_handler()
73 hmask = kvm_riscv_vcpu_unpriv_read(vcpu, false, cp->a0, utrap); in kvm_sbi_ext_v01_handler()
79 if (cp->a7 == SBI_EXT_0_1_REMOTE_FENCE_I) in kvm_sbi_ext_v01_handler()
81 else if (cp->a7 == SBI_EXT_0_1_REMOTE_SFENCE_VMA) { in kvm_sbi_ext_v01_handler()
[all …]
/linux/net/bluetooth/
H A Dhci_sync.c901 struct hci_cp_write_eir cp; in hci_update_eir_sync() local
917 memset(&cp, 0, sizeof(cp)); in hci_update_eir_sync()
919 eir_create(hdev, cp.data); in hci_update_eir_sync()
921 if (memcmp(cp.data, hdev->eir, sizeof(cp.data)) == 0) in hci_update_eir_sync()
924 memcpy(hdev->eir, cp.data, sizeof(cp.data)); in hci_update_eir_sync()
926 return __hci_cmd_sync_status(hdev, HCI_OP_WRITE_EIR, sizeof(cp), &cp, in hci_update_eir_sync()
1148 struct hci_cp_le_set_ext_adv_enable *cp; in hci_disable_ext_adv_instance_sync() local
1150 u8 data[sizeof(*cp) + sizeof(*set) * 1]; in hci_disable_ext_adv_instance_sync()
1167 cp = (void *)data; in hci_disable_ext_adv_instance_sync()
1168 set = (void *)cp->data; in hci_disable_ext_adv_instance_sync()
[all …]
H A Dmgmt.c957 struct mgmt_mode *cp = cmd->param; in mgmt_get_adv_discov_flags() local
958 if (cp->val == 0x01) in mgmt_get_adv_discov_flags()
960 else if (cp->val == 0x02) in mgmt_get_adv_discov_flags()
981 struct mgmt_mode *cp = cmd->param; in mgmt_get_connectable() local
983 return cp->val; in mgmt_get_connectable()
1323 struct mgmt_mode *cp; in mgmt_set_powered_complete() local
1329 cp = cmd->param; in mgmt_set_powered_complete()
1334 if (cp->val) { in mgmt_set_powered_complete()
1346 if (cp->val) in mgmt_set_powered_complete()
1359 struct mgmt_mode cp; in set_powered_sync() local
[all …]
/linux/drivers/tty/vt/
H A Ducs.c25 u16 cp = *(u16 *)key; in interval16_cmp() local
28 if (cp < entry->first) in interval16_cmp()
30 if (cp > entry->last) in interval16_cmp()
37 u32 cp = *(u32 *)key; in interval32_cmp() local
40 if (cp < entry->first) in interval32_cmp()
42 if (cp > entry->last) in interval32_cmp()
47 static bool cp_in_range16(u16 cp, const struct ucs_interval16 *ranges, size_t size) in cp_in_range16() argument
49 if (cp < ranges[0].first || cp > ranges[size - 1].last) in cp_in_range16()
52 return __inline_bsearch(&cp, ranges, size, sizeof(*ranges), in cp_in_range16()
56 static bool cp_in_range32(u32 cp, const struct ucs_interval32 *ranges, size_t size) in cp_in_range32() argument
[all …]
/linux/drivers/accessibility/speakup/
H A Dkobjects.c34 char *cp; in chars_chartab_show() local
49 cp = "B_CTL"; in chars_chartab_show()
51 cp = "WDLM"; in chars_chartab_show()
53 cp = "A_PUNC"; in chars_chartab_show()
55 cp = "PUNC"; in chars_chartab_show()
57 cp = "NUM"; in chars_chartab_show()
59 cp = "A_CAP"; in chars_chartab_show()
61 cp = "ALPHA"; in chars_chartab_show()
63 cp = "B_CAPSYM"; in chars_chartab_show()
65 cp = "B_SYM"; in chars_chartab_show()
[all …]
/linux/include/sound/
H A Dseq_midi_emul.h134 #define SNDRV_GM_BANK_SELECT(cp) (((cp)->control[0]<<7)|((cp)->control[32])) argument
135 #define SNDRV_GM_MODULATION_WHEEL(cp) (((cp)->control[1]<<7)|((cp)->control[33])) argument
136 #define SNDRV_GM_BREATH(cp) (((cp)->control[2]<<7)|((cp)->control[34])) argument
137 #define SNDRV_GM_FOOT_PEDAL(cp) (((cp)->control[4]<<7)|((cp)->control[36])) argument
138 #define SNDRV_GM_PORTAMENTO_TIME(cp) (((cp)->control[5]<<7)|((cp)->control[37])) argument
139 #define SNDRV_GM_DATA_ENTRY(cp) (((cp)->control[6]<<7)|((cp)->control[38])) argument
140 #define SNDRV_GM_VOLUME(cp) (((cp)->control[7]<<7)|((cp)->control[39])) argument
141 #define SNDRV_GM_BALANCE(cp) (((cp)->control[8]<<7)|((cp)->control[40])) argument
142 #define SNDRV_GM_PAN(cp) (((cp)->control[10]<<7)|((cp)->control[42])) argument
143 #define SNDRV_GM_EXPRESSION(cp) (((cp)->control[11]<<7)|((cp)->control[43])) argument
/linux/drivers/s390/cio/
H A Dvfio_ccw_cp.c318 static struct ccwchain *ccwchain_alloc(struct channel_program *cp, int len) in ccwchain_alloc() argument
334 list_add_tail(&chain->next, &cp->ccwchain_list); in ccwchain_alloc()
366 * @cp: channel_program on which to perform the operation
376 static int ccwchain_calc_length(u64 iova, struct channel_program *cp) in ccwchain_calc_length() argument
378 struct ccw1 *ccw = cp->guest_cp; in ccwchain_calc_length()
404 static int tic_target_chain_exists(struct ccw1 *tic, struct channel_program *cp) in tic_target_chain_exists() argument
409 list_for_each_entry(chain, &cp->ccwchain_list, next) { in tic_target_chain_exists()
419 struct channel_program *cp);
421 static int ccwchain_handle_ccw(dma32_t cda, struct channel_program *cp) in ccwchain_handle_ccw() argument
424 &container_of(cp, struct vfio_ccw_private, cp)->vdev; in ccwchain_handle_ccw()
[all …]
/linux/drivers/scsi/sym53c8xx_2/
H A Dsym_hipd.c46 static void sym_complete_error (struct sym_hcb *np, struct sym_ccb *cp);
47 static void sym_complete_ok (struct sym_hcb *np, struct sym_ccb *cp);
48 static int sym_compute_residual(struct sym_hcb *np, struct sym_ccb *cp);
60 static void sym_print_msg(struct sym_ccb *cp, char *label, u_char *msg) in sym_print_msg() argument
62 sym_print_addr(cp->cmd, "%s: ", label); in sym_print_msg()
1406 static int sym_prepare_nego(struct sym_hcb *np, struct sym_ccb *cp, u_char *msgptr) argument
1408 struct sym_tcb *tp = &np->target[cp->target];
1449 cp->nego_status = nego;
1452 tp->nego_cp = cp; /* Keep track a nego will be performed */
1454 sym_print_nego_msg(np, cp->target,
[all …]

12345678910>>...33