/linux/net/rxrpc/ |
H A D | call_object.c | 2 /* RxRPC individual remote procedure call handling 45 void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what) in rxrpc_poke_call() argument 47 struct rxrpc_local *local = call->local; in rxrpc_poke_call() 50 if (!test_bit(RXRPC_CALL_DISCONNECTED, &call->flags)) { in rxrpc_poke_call() 52 busy = !list_empty(&call->attend_link); in rxrpc_poke_call() 53 trace_rxrpc_poke_call(call, busy, what); in rxrpc_poke_call() 54 if (!busy && !rxrpc_try_get_call(call, rxrpc_call_get_poke)) in rxrpc_poke_call() 57 list_add_tail(&call->attend_link, &local->call_attend_q); in rxrpc_poke_call() 67 struct rxrpc_call *call = timer_container_of(call, t, timer); in rxrpc_call_timer_expired() local 69 _enter("%d", call->debug_id); in rxrpc_call_timer_expired() [all …]
|
H A D | call_event.c | 23 void rxrpc_propose_ping(struct rxrpc_call *call, u32 serial, in rxrpc_propose_ping() argument 30 trace_rxrpc_propose_ack(call, why, RXRPC_ACK_PING, serial); in rxrpc_propose_ping() 31 if (ktime_before(ping_at, call->ping_at)) { in rxrpc_propose_ping() 32 call->ping_at = ping_at; in rxrpc_propose_ping() 33 trace_rxrpc_timer_set(call, delay, rxrpc_timer_trace_ping); in rxrpc_propose_ping() 40 void rxrpc_propose_delay_ACK(struct rxrpc_call *call, rxrpc_serial_t serial, in rxrpc_propose_delay_ACK() argument 45 trace_rxrpc_propose_ack(call, why, RXRPC_ACK_DELAY, serial); in rxrpc_propose_delay_ACK() 47 if (call->srtt_us) in rxrpc_propose_delay_ACK() 48 delay = (call->srtt_us >> 3) * NSEC_PER_USEC; in rxrpc_propose_delay_ACK() 51 ktime_add_ms(delay, call->tx_backoff); in rxrpc_propose_delay_ACK() [all …]
|
H A D | input.c | 23 static void rxrpc_proto_abort(struct rxrpc_call *call, rxrpc_seq_t seq, in rxrpc_proto_abort() argument 26 rxrpc_abort_call(call, seq, RX_PROTOCOL_ERROR, -EBADMSG, why); in rxrpc_proto_abort() 32 static void rxrpc_congestion_management(struct rxrpc_call *call, in rxrpc_congestion_management() argument 36 summary->in_flight = rxrpc_tx_in_flight(call); in rxrpc_congestion_management() 38 if (test_and_clear_bit(RXRPC_CALL_RETRANS_TIMEOUT, &call->flags)) { in rxrpc_congestion_management() 40 call->cong_ssthresh = umax(summary->in_flight / 2, 2); in rxrpc_congestion_management() 41 call->cong_cwnd = 1; in rxrpc_congestion_management() 42 if (call->cong_cwnd >= call->cong_ssthresh && in rxrpc_congestion_management() 43 call->cong_ca_state == RXRPC_CA_SLOW_START) { in rxrpc_congestion_management() 44 call->cong_ca_state = RXRPC_CA_CONGEST_AVOIDANCE; in rxrpc_congestion_management() [all …]
|
H A D | sendmsg.c | 23 bool rxrpc_propose_abort(struct rxrpc_call *call, s32 abort_code, int error, in rxrpc_propose_abort() argument 26 _enter("{%d},%d,%d,%u", call->debug_id, abort_code, error, why); in rxrpc_propose_abort() 28 if (!call->send_abort && !rxrpc_call_is_complete(call)) { in rxrpc_propose_abort() 29 call->send_abort_why = why; in rxrpc_propose_abort() 30 call->send_abort_err = error; in rxrpc_propose_abort() 31 call->send_abort_seq = 0; in rxrpc_propose_abort() 32 trace_rxrpc_abort_call(call, abort_code); in rxrpc_propose_abort() 34 smp_store_release(&call->send_abort, abort_code); in rxrpc_propose_abort() 35 rxrpc_poke_call(call, rxrpc_call_poke_abor in rxrpc_propose_abort() 46 rxrpc_wait_to_be_connected(struct rxrpc_call * call,long * timeo) rxrpc_wait_to_be_connected() argument 95 rxrpc_check_tx_space(struct rxrpc_call * call,rxrpc_seq_t * _tx_win) rxrpc_check_tx_space() argument 108 rxrpc_wait_for_tx_window_intr(struct rxrpc_sock * rx,struct rxrpc_call * call,long * timeo) rxrpc_wait_for_tx_window_intr() argument 132 rxrpc_wait_for_tx_window_waitall(struct rxrpc_sock * rx,struct rxrpc_call * call) rxrpc_wait_for_tx_window_waitall() argument 172 rxrpc_wait_for_tx_window_nonintr(struct rxrpc_sock * rx,struct rxrpc_call * call,long * timeo) rxrpc_wait_for_tx_window_nonintr() argument 193 rxrpc_wait_for_tx_window(struct rxrpc_sock * rx,struct rxrpc_call * call,long * timeo,bool waitall) rxrpc_wait_for_tx_window() argument 229 rxrpc_notify_end_tx(struct rxrpc_sock * rx,struct rxrpc_call * call,rxrpc_notify_end_tx_t notify_end_tx) rxrpc_notify_end_tx() argument 241 rxrpc_queue_packet(struct rxrpc_sock * rx,struct rxrpc_call * call,struct rxrpc_txbuf * txb,rxrpc_notify_end_tx_t notify_end_tx) rxrpc_queue_packet() argument 281 rxrpc_alloc_txqueue(struct sock * sk,struct rxrpc_call * call) rxrpc_alloc_txqueue() argument 321 rxrpc_send_data(struct rxrpc_sock * rx,struct rxrpc_call * call,struct msghdr * msg,size_t len,rxrpc_notify_end_tx_t notify_end_tx,bool * _dropped_lock) rxrpc_send_data() argument 615 struct rxrpc_call *call; rxrpc_new_client_call_for_sendmsg() local 662 struct rxrpc_call *call; rxrpc_do_sendmsg() local 798 rxrpc_kernel_send_data(struct socket * sock,struct rxrpc_call * call,struct msghdr * msg,size_t len,rxrpc_notify_end_tx_t notify_end_tx) rxrpc_kernel_send_data() argument 835 rxrpc_kernel_abort_call(struct socket * sock,struct rxrpc_call * call,u32 abort_code,int error,enum rxrpc_abort_reason why) rxrpc_kernel_abort_call() argument 861 rxrpc_kernel_set_tx_length(struct socket * sock,struct rxrpc_call * call,s64 tx_total_len) rxrpc_kernel_set_tx_length() argument [all...] |
H A D | rxperf.c | 66 int (*deliver)(struct rxperf_call *call); 75 static int rxperf_deliver_param_block(struct rxperf_call *call); 76 static int rxperf_deliver_request(struct rxperf_call *call); 77 static int rxperf_process_call(struct rxperf_call *call); 83 static inline void rxperf_set_call_state(struct rxperf_call *call, in rxperf_set_call_state() 86 call->state = to; in rxperf_set_call_state() 89 static inline void rxperf_set_call_complete(struct rxperf_call *call, in rxperf_set_call_complete() 92 if (call->state != RXPERF_CALL_COMPLETE) { in rxperf_set_call_complete() 93 call->abort_code = remote_abort; in rxperf_set_call_complete() 94 call in rxperf_set_call_complete() 82 rxperf_set_call_state(struct rxperf_call * call,enum rxperf_call_state to) rxperf_set_call_state() argument 88 rxperf_set_call_complete(struct rxperf_call * call,int error,s32 remote_abort) rxperf_set_call_complete() argument 110 rxperf_queue_call_work(struct rxperf_call * call) rxperf_queue_call_work() argument 118 struct rxperf_call *call = (struct rxperf_call *)call_user_ID; rxperf_notify_rx() local 126 struct rxperf_call *call = (struct rxperf_call *)user_call_ID; rxperf_rx_attach() local 144 struct rxperf_call *call; rxperf_charge_preallocation() local 245 rxperf_log_error(struct rxperf_call * call,s32 remote_abort) rxperf_log_error() argument 277 struct rxperf_call *call = container_of(work, struct rxperf_call, work); rxperf_deliver_to_call() local 354 rxperf_extract_data(struct rxperf_call * call,bool want_more) rxperf_extract_data() argument 388 rxperf_deliver_param_block(struct rxperf_call * call) rxperf_deliver_param_block() argument 436 rxperf_deliver_request(struct rxperf_call * call) rxperf_deliver_request() argument 506 rxperf_process_call(struct rxperf_call * call) rxperf_process_call() argument [all...] |
H A D | recvmsg.c | 20 * Post a call for attention by the socket or kernel service. Further 23 void rxrpc_notify_socket(struct rxrpc_call *call) in rxrpc_notify_socket() argument 28 _enter("%d", call->debug_id); in rxrpc_notify_socket() 30 if (!list_empty(&call->recvmsg_link)) in rxrpc_notify_socket() 32 if (test_bit(RXRPC_CALL_RELEASED, &call->flags)) { in rxrpc_notify_socket() 33 rxrpc_see_call(call, rxrpc_call_see_notify_released); in rxrpc_notify_socket() 39 rx = rcu_dereference(call->socket); in rxrpc_notify_socket() 42 if (call->notify_rx) { in rxrpc_notify_socket() 43 spin_lock_irq(&call->notify_lock); in rxrpc_notify_socket() 44 call->notify_rx(sk, call, call->user_call_ID); in rxrpc_notify_socket() [all …]
|
H A D | output.c | 48 static void rxrpc_tx_backoff(struct rxrpc_call *call, int ret) in rxrpc_tx_backoff() argument 51 if (call->tx_backoff < 1000) in rxrpc_tx_backoff() 52 call->tx_backoff += 100; in rxrpc_tx_backoff() 54 call->tx_backoff = 0; in rxrpc_tx_backoff() 60 * lets the far side know we're still interested in this call and helps keep 66 static void rxrpc_set_keepalive(struct rxrpc_call *call, ktime_t now) in rxrpc_set_keepalive() argument 68 ktime_t delay = ms_to_ktime(READ_ONCE(call->next_rx_timo) / 6); in rxrpc_set_keepalive() 70 call->keepalive_at = ktime_add(ktime_get_real(), delay); in rxrpc_set_keepalive() 71 trace_rxrpc_timer_set(call, delay, rxrpc_timer_trace_keepalive); in rxrpc_set_keepalive() 77 static int rxrpc_alloc_ack(struct rxrpc_call *call, size_t sack_size) in rxrpc_alloc_ack() argument [all …]
|
H A D | call_state.c | 2 /* Call state changing functions. 11 * Transition a call to the complete state. 13 bool rxrpc_set_call_completion(struct rxrpc_call *call, in rxrpc_set_call_completion() argument 18 if (__rxrpc_call_state(call) == RXRPC_CALL_COMPLETE) in rxrpc_set_call_completion() 21 call->abort_code = abort_code; in rxrpc_set_call_completion() 22 call->error = error; in rxrpc_set_call_completion() 23 call->completion = compl; in rxrpc_set_call_completion() 25 rxrpc_set_call_state(call, RXRPC_CALL_COMPLETE); in rxrpc_set_call_completion() 26 trace_rxrpc_call_complete(call); in rxrpc_set_call_completion() 27 wake_up(&call->waitq); in rxrpc_set_call_completion() [all …]
|
H A D | call_accept.c | 2 /* incoming call handling 25 static void rxrpc_dummy_notify(struct sock *sk, struct rxrpc_call *call, in rxrpc_dummy_notify() argument 31 * Preallocate a single service call, connection and peer and, if possible, 40 struct rxrpc_call *call, *xcall; in rxrpc_service_prealloc_one() local 97 call = rxrpc_alloc_call(rx, gfp, debug_id); in rxrpc_service_prealloc_one() 98 if (!call) in rxrpc_service_prealloc_one() 100 call->flags |= (1 << RXRPC_CALL_IS_SERVICE); in rxrpc_service_prealloc_one() 101 rxrpc_set_call_state(call, RXRPC_CALL_SERVER_PREALLOC); in rxrpc_service_prealloc_one() 102 __set_bit(RXRPC_CALL_EV_INITIAL_PING, &call->events); in rxrpc_service_prealloc_one() 104 trace_rxrpc_call(call->debug_id, refcount_read(&call->ref), in rxrpc_service_prealloc_one() [all …]
|
H A D | conn_client.c | 8 * call so as to handle retransmitted DATA packets in case the server didn't 15 * or a call ID counter overflows. 73 static struct rxrpc_bundle *rxrpc_alloc_bundle(struct rxrpc_call *call, in rxrpc_alloc_bundle() argument 81 bundle->local = call->local; in rxrpc_alloc_bundle() 82 bundle->peer = rxrpc_get_peer(call->peer, rxrpc_peer_get_bundle); in rxrpc_alloc_bundle() 83 bundle->key = key_get(call->key); in rxrpc_alloc_bundle() 84 bundle->security = call->security; in rxrpc_alloc_bundle() 85 bundle->exclusive = test_bit(RXRPC_CALL_EXCLUSIVE, &call->flags); in rxrpc_alloc_bundle() 86 bundle->upgrade = test_bit(RXRPC_CALL_UPGRADE, &call->flags); in rxrpc_alloc_bundle() 87 bundle->service_id = call->dest_srx.srx_service; in rxrpc_alloc_bundle() [all …]
|
/linux/fs/afs/ |
H A D | vlclient.c | 15 * Deliver reply data to a VL.GetEntryByNameU call. 17 static int afs_deliver_vl_get_entry_by_name_u(struct afs_call *call) in afs_deliver_vl_get_entry_by_name_u() argument 26 ret = afs_transfer_reply(call); in afs_deliver_vl_get_entry_by_name_u() 31 uvldb = call->buffer; in afs_deliver_vl_get_entry_by_name_u() 32 entry = call->ret_vldb; in afs_deliver_vl_get_entry_by_name_u() 115 struct afs_call *call; in afs_vl_get_entry_by_name_u() local 129 call = afs_alloc_flat_call(net, &afs_RXVLGetEntryByNameU, reqsz, in afs_vl_get_entry_by_name_u() 131 if (!call) { in afs_vl_get_entry_by_name_u() 136 call->key = vc->key; in afs_vl_get_entry_by_name_u() 137 call->ret_vldb = entry; in afs_vl_get_entry_by_name_u() [all …]
|
H A D | yfsclient.c | 138 static void yfs_check_req(struct afs_call *call, __be32 *bp) in yfs_check_req() argument 140 size_t len = (void *)bp - call->request; in yfs_check_req() 142 if (len > call->request_size) in yfs_check_req() 144 call->type->name, len, call->request_size); in yfs_check_req() 145 else if (len < call->request_size) in yfs_check_req() 147 call->type->name, len, call->request_size); in yfs_check_req() 174 struct afs_call *call, in xdr_decode_YFSFetchStatus() argument 220 afs_protocol_error(call, afs_eproto_bad_status); in xdr_decode_YFSFetchStatus() 228 struct afs_call *call, in xdr_decode_YFSCallBack() argument 235 cb_expiry = ktime_add(call->issue_time, xdr_to_u64(x->expiration_time) * 100); in xdr_decode_YFSCallBack() [all …]
|
/linux/include/trace/events/ |
H A D | rxrpc.h | 52 EM(rxkad_abort_resp_call_ctr, "rxkad-resp-call-ctr") \ 53 EM(rxkad_abort_resp_call_state, "rxkad-resp-call-state") \ 86 EM(rxgk_abort_resp_call_ctr, "rxgk-resp-call-ctr") \ 87 EM(rxgk_abort_resp_call_state, "rxgk-resp-call-state") \ 105 EM(rxrpc_abort_call_improper_term, "call-improper-term") \ 106 EM(rxrpc_abort_call_reset, "call-reset") \ 107 EM(rxrpc_abort_call_sendmsg, "call-sendmsg") \ 108 EM(rxrpc_abort_call_sock_release, "call-sock-rel") \ 109 EM(rxrpc_abort_call_sock_release_tba, "call-sock-rel-tba") \ 110 EM(rxrpc_abort_call_timeout, "call-timeout") \ [all …]
|
/linux/include/trace/ |
H A D | trace_events.h | 8 * struct trace_event_raw_<call> { 114 * struct trace_event_data_offsets_<call> { 128 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument 129 struct trace_event_data_offsets_##call { \ 157 * trace_raw_output_<call>(struct trace_iterator *iter, int flags) 160 * struct trace_event_raw_<call> *field; <-- defined in stage 1 169 * if (entry->type != event_<call>->event.type) { 177 * return trace_output_call(iter, <call>, <TP_printk> "\n"); 203 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument 205 trace_raw_output_##call(struct trace_iterator *iter, int flags, \ [all …]
|
H A D | trace_custom_events.h | 62 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument 63 struct trace_custom_event_data_offsets_##call { \ 77 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument 79 trace_custom_raw_output_##call(struct trace_iterator *iter, int flags, \ 84 struct trace_custom_event_raw_##call *field; \ 97 static struct trace_event_functions trace_custom_event_type_funcs_##call = { \ 98 .trace = trace_custom_raw_output_##call, \ 108 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, func, print) \ argument 109 static struct trace_event_fields trace_custom_event_fields_##call[] = { \ 120 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument [all …]
|
H A D | bpf_probe.h | 45 #define __BPF_DECLARE_TRACE(call, proto, args) \ argument 47 __bpf_trace_##call(void *__data, proto) \ 53 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument 54 __BPF_DECLARE_TRACE(call, PARAMS(proto), PARAMS(args)) 56 #define __BPF_DECLARE_TRACE_SYSCALL(call, proto, args) \ argument 58 __bpf_trace_##call(void *__data, proto) \ 67 #define DECLARE_EVENT_SYSCALL_CLASS(call, proto, args, tstruct, assign, print) \ argument 68 __BPF_DECLARE_TRACE_SYSCALL(call, PARAMS(proto), PARAMS(args)) 75 #define __DEFINE_EVENT(template, call, proto, args, size) \ argument 76 static inline void bpf_test_probe_##call(void) \ [all …]
|
/linux/include/asm-generic/ |
H A D | syscall.h | 3 * Access to user system call parameters and results 23 * syscall_get_nr - find what system call a task is executing 27 * If @task is executing a system call or is at system call 28 * tracing about to attempt one, returns the system call number. 29 * If @task is not executing a system call, i.e. it's blocked 33 * system call number can be meaningful. If the actual arch value 36 * It's only valid to call this when @task is known to be blocked. 41 * syscall_set_nr - change the system call a task is executing 44 * @nr: system call number 46 * Changes the system call number @task is about to execute. [all …]
|
/linux/tools/ |
H A D | Makefile | 67 $(call descend,power/$@) 70 $(call descend,power/$@) 73 $(call descend,$@) 76 $(call descend,$@) 79 $(call descend,lib/api) 82 $(call descend,include/nolibc) 85 $(call descend,include/nolibc,$(patsubst nolibc_%,%,$@)) 96 $(call descend,sched_ext) 99 $(call descend,testing/$@) 102 $(call descen [all...] |
/linux/include/linux/firmware/intel/ |
H A D | stratix10-smc.h | 13 * This file defines the Secure Monitor Call (SMC) message protocol used for 29 * FAST call executes atomic operations, returns when the requested operation 31 * STD call starts a operation which can be preempted by a non-secure 32 * interrupt. The call can return before the requested operation has 51 * Return values in INTEL_SIP_SMC_* call 81 * Sync call used by service driver at EL1 to request the FPGA in EL3 to 84 * Call register usage: 101 * Async call used by service driver at EL1 to provide FPGA configuration data 104 * Call register usage: 127 * Sync call used by service driver at EL1 to track the completed write [all …]
|
/linux/tools/perf/util/ |
H A D | thread-stack.h | 3 * thread-stack.h: Synthesize a thread's stack using call / return events 24 * Call/Return flags. 26 * CALL_RETURN_NO_CALL: 'return' but no matching 'call' 27 * CALL_RETURN_NO_RETURN: 'call' but no matching 'return' 28 * CALL_RETURN_NON_CALL: a branch but not a 'call' to the start of a different 38 * struct call_return - paired call/return information. 39 * @thread: thread in which call/return occurred 40 * @comm: comm in which call/return occurred 41 * @cp: call path 42 * @call_time: timestamp of call (if known) [all …]
|
/linux/scripts/ |
H A D | Makefile.compiler | 4 # Usage: CROSS_COMPILE := $(call cc-cross-prefix, m68k-linux-gnu- m68k-linux-) 19 # Usage: option = $(call try-run, $(CC)...-o "$$TMP",option-ok,otherwise) 32 # Usage: aflags-y += $(call as-option,-Wa$(comma)-isa=foo,) 34 as-option = $(call try-run,\ 38 # Usage: aflags-y += $(call as-instr,instr,option1,option2) 40 as-instr = $(call try-run,\ 44 # Usage: MY_CFLAGS += $(call __cc-option,$(CC),$(MY_CFLAGS),-march=winchip-c6,-march=i586) 45 __cc-option = $(call try-run,\ 49 # Usage: cflags-y += $(call cc-option,-march=winchip-c6,-march=i586) 51 cc-option = $(call __cc-option, $(CC),\ [all …]
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | verifier_spin_lock.c | 32 call %[bpf_map_lookup_elem]; \ in spin_lock_test1_success() 38 call %[bpf_spin_lock]; \ in spin_lock_test1_success() 42 call %[bpf_spin_unlock]; \ in spin_lock_test1_success() 65 call %[bpf_map_lookup_elem]; \ in lock_test2_direct_ld_st() 71 call %[bpf_spin_lock]; \ in lock_test2_direct_ld_st() 75 call %[bpf_spin_unlock]; \ in lock_test2_direct_ld_st() 99 call %[bpf_map_lookup_elem]; \ in __flag() 105 call %[bpf_spin_lock]; \ in __flag() 109 call %[bpf_spin_unlock]; \ in __flag() 133 call %[bpf_map_lookup_elem]; \ in __flag() [all …]
|
H A D | verifier_subreg.c | 27 call %[bpf_get_prandom_u32]; \ in add32_reg_zero_extend_check() 44 call %[bpf_get_prandom_u32]; \ in add32_imm_zero_extend_check() 56 call %[bpf_get_prandom_u32]; \ in add32_imm_zero_extend_check() 74 call %[bpf_get_prandom_u32]; \ in sub32_reg_zero_extend_check() 91 call %[bpf_get_prandom_u32]; \ in sub32_imm_zero_extend_check() 97 call %[bpf_get_prandom_u32]; \ in sub32_imm_zero_extend_check() 115 call %[bpf_get_prandom_u32]; \ in mul32_reg_zero_extend_check() 132 call %[bpf_get_prandom_u32]; \ in mul32_imm_zero_extend_check() 138 call %[bpf_get_prandom_u32]; \ in mul32_imm_zero_extend_check() 156 call %[bpf_get_prandom_u32]; \ in div32_reg_zero_extend_check() [all …]
|
H A D | verifier_ref_tracking.c | 24 "call %[" #func "];" 101 call %[bpf_tail_call]; \ in dummy_prog_loop1_tc() 233 call %[bpf_lookup_user_key]; \ in acquire_release_user_key_reference() 236 call %[bpf_key_put]; \ in acquire_release_user_key_reference() 252 call %[bpf_lookup_system_key]; \ in acquire_release_system_key_reference() 255 call %[bpf_key_put]; \ in acquire_release_system_key_reference() 272 call %[bpf_lookup_user_key]; \ in user_key_reference_without_check() 274 call %[bpf_key_put]; \ in user_key_reference_without_check() 290 call %[bpf_lookup_system_key]; \ in system_key_reference_without_check() 292 call %[bpf_key_put]; \ in system_key_reference_without_check() [all …]
|
/linux/tools/perf/ |
H A D | Makefile.config | 64 $(call detected_var,SRCARCH) 78 $(call detected,CONFIG_X86) 85 $(call detected,CONFIG_X86_64) 152 $(foreach libunwind_arch,$(LIBUNWIND_ARCHS),$(call libunwind_arch_set_flags,$(libunwind_arch))) 236 ifeq ($(call get-executable,$(PKG_CONFIG)),) 247 ifeq ($(call get-executable,$(FLEX)),) 251 ifeq ($(call get-executable,$(BISON)),) 284 $(call detected_var,PARSER_DEBUG_BISON) 285 $(call detected_var,PARSER_DEBUG_FLEX) 302 PYTHON_AUTO := $(if $(call get-executable,python2-config),python2-config,$(PYTHON_AUTO)) [all …]
|