Home
last modified time | relevance | path

Searched full:call (Results 1 – 25 of 5637) sorted by relevance

12345678910>>...226

/linux-6.14.4/fs/afs/
Drxrpc.c29 /* asynchronous incoming call initial processing */
135 * Allocate a call.
141 struct afs_call *call; in afs_alloc_call() local
144 call = kzalloc(sizeof(*call), gfp); in afs_alloc_call()
145 if (!call) in afs_alloc_call()
148 call->type = type; in afs_alloc_call()
149 call->net = net; in afs_alloc_call()
150 call->debug_id = atomic_inc_return(&rxrpc_debug_id); in afs_alloc_call()
151 refcount_set(&call->ref, 1); in afs_alloc_call()
152 INIT_WORK(&call->async_work, type->async_rx ?: afs_process_async_call); in afs_alloc_call()
[all …]
Dcmservice.c105 * route an incoming cache manager call
108 bool afs_cm_incoming_call(struct afs_call *call) in afs_cm_incoming_call() argument
110 _enter("{%u, CB.OP %u}", call->service_id, call->operation_ID); in afs_cm_incoming_call()
112 switch (call->operation_ID) { in afs_cm_incoming_call()
114 call->type = &afs_SRXCBCallBack; in afs_cm_incoming_call()
117 call->type = &afs_SRXCBInitCallBackState; in afs_cm_incoming_call()
120 call->type = &afs_SRXCBInitCallBackState3; in afs_cm_incoming_call()
123 call->type = &afs_SRXCBProbe; in afs_cm_incoming_call()
126 call->type = &afs_SRXCBProbeUuid; in afs_cm_incoming_call()
129 call->type = &afs_SRXCBTellMeAboutYourself; in afs_cm_incoming_call()
[all …]
Dvlclient.c15 * Deliver reply data to a VL.GetEntryByNameU call.
17 static int afs_deliver_vl_get_entry_by_name_u(struct afs_call *call) in afs_deliver_vl_get_entry_by_name_u() argument
26 ret = afs_transfer_reply(call); in afs_deliver_vl_get_entry_by_name_u()
31 uvldb = call->buffer; in afs_deliver_vl_get_entry_by_name_u()
32 entry = call->ret_vldb; in afs_deliver_vl_get_entry_by_name_u()
115 struct afs_call *call; in afs_vl_get_entry_by_name_u() local
129 call = afs_alloc_flat_call(net, &afs_RXVLGetEntryByNameU, reqsz, in afs_vl_get_entry_by_name_u()
131 if (!call) { in afs_vl_get_entry_by_name_u()
136 call->key = vc->key; in afs_vl_get_entry_by_name_u()
137 call->ret_vldb = entry; in afs_vl_get_entry_by_name_u()
[all …]
Dfsclient.c55 struct afs_call *call, in xdr_decode_AFSFetchStatus() argument
60 bool inline_error = (call->operation_ID == afs_FS_InlineBulkStatus); in xdr_decode_AFSFetchStatus()
128 afs_protocol_error(call, afs_eproto_bad_status); in xdr_decode_AFSFetchStatus()
132 static time64_t xdr_decode_expiry(struct afs_call *call, u32 expiry) in xdr_decode_expiry() argument
134 return ktime_divns(call->issue_time, NSEC_PER_SEC) + expiry; in xdr_decode_expiry()
138 struct afs_call *call, in xdr_decode_AFSCallBack() argument
145 cb->expires_at = xdr_decode_expiry(call, ntohl(*bp++)); in xdr_decode_AFSCallBack()
238 static int afs_deliver_fs_fetch_status(struct afs_call *call) in afs_deliver_fs_fetch_status() argument
240 struct afs_operation *op = call->op; in afs_deliver_fs_fetch_status()
245 ret = afs_transfer_reply(call); in afs_deliver_fs_fetch_status()
[all …]
Dyfsclient.c138 static void yfs_check_req(struct afs_call *call, __be32 *bp) in yfs_check_req() argument
140 size_t len = (void *)bp - call->request; in yfs_check_req()
142 if (len > call->request_size) in yfs_check_req()
144 call->type->name, len, call->request_size); in yfs_check_req()
145 else if (len < call->request_size) in yfs_check_req()
147 call->type->name, len, call->request_size); in yfs_check_req()
174 struct afs_call *call, in xdr_decode_YFSFetchStatus() argument
220 afs_protocol_error(call, afs_eproto_bad_status); in xdr_decode_YFSFetchStatus()
228 struct afs_call *call, in xdr_decode_YFSCallBack() argument
235 cb_expiry = ktime_add(call->issue_time, xdr_to_u64(x->expiration_time) * 100); in xdr_decode_YFSCallBack()
[all …]
/linux-6.14.4/net/rxrpc/
Dcall_object.c2 /* RxRPC individual remote procedure call handling
45 void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what) in rxrpc_poke_call() argument
47 struct rxrpc_local *local = call->local; in rxrpc_poke_call()
50 if (!test_bit(RXRPC_CALL_DISCONNECTED, &call->flags)) { in rxrpc_poke_call()
52 busy = !list_empty(&call->attend_link); in rxrpc_poke_call()
53 trace_rxrpc_poke_call(call, busy, what); in rxrpc_poke_call()
54 if (!busy && !rxrpc_try_get_call(call, rxrpc_call_get_poke)) in rxrpc_poke_call()
57 list_add_tail(&call->attend_link, &local->call_attend_q); in rxrpc_poke_call()
67 struct rxrpc_call *call = from_timer(call, t, timer); in rxrpc_call_timer_expired() local
69 _enter("%d", call->debug_id); in rxrpc_call_timer_expired()
[all …]
Dcall_event.c23 void rxrpc_propose_ping(struct rxrpc_call *call, u32 serial, in rxrpc_propose_ping() argument
30 trace_rxrpc_propose_ack(call, why, RXRPC_ACK_PING, serial); in rxrpc_propose_ping()
31 if (ktime_before(ping_at, call->ping_at)) { in rxrpc_propose_ping()
32 call->ping_at = ping_at; in rxrpc_propose_ping()
33 trace_rxrpc_timer_set(call, delay, rxrpc_timer_trace_ping); in rxrpc_propose_ping()
40 void rxrpc_propose_delay_ACK(struct rxrpc_call *call, rxrpc_serial_t serial, in rxrpc_propose_delay_ACK() argument
45 trace_rxrpc_propose_ack(call, why, RXRPC_ACK_DELAY, serial); in rxrpc_propose_delay_ACK()
47 if (call->srtt_us) in rxrpc_propose_delay_ACK()
48 delay = (call->srtt_us >> 3) * NSEC_PER_USEC; in rxrpc_propose_delay_ACK()
51 ktime_add_ms(delay, call->tx_backoff); in rxrpc_propose_delay_ACK()
[all …]
Dinput.c23 static void rxrpc_proto_abort(struct rxrpc_call *call, rxrpc_seq_t seq, in rxrpc_proto_abort() argument
26 rxrpc_abort_call(call, seq, RX_PROTOCOL_ERROR, -EBADMSG, why); in rxrpc_proto_abort()
32 static void rxrpc_congestion_management(struct rxrpc_call *call, in rxrpc_congestion_management() argument
36 summary->in_flight = rxrpc_tx_in_flight(call); in rxrpc_congestion_management()
38 if (test_and_clear_bit(RXRPC_CALL_RETRANS_TIMEOUT, &call->flags)) { in rxrpc_congestion_management()
40 call->cong_ssthresh = umax(summary->in_flight / 2, 2); in rxrpc_congestion_management()
41 call->cong_cwnd = 1; in rxrpc_congestion_management()
42 if (call->cong_cwnd >= call->cong_ssthresh && in rxrpc_congestion_management()
43 call->cong_ca_state == RXRPC_CA_SLOW_START) { in rxrpc_congestion_management()
44 call->cong_ca_state = RXRPC_CA_CONGEST_AVOIDANCE; in rxrpc_congestion_management()
[all …]
Dinput_rack.c23 static void rxrpc_rack_mark_lost(struct rxrpc_call *call, in rxrpc_rack_mark_lost() argument
28 call->tx_nr_resent--; in rxrpc_rack_mark_lost()
30 call->tx_nr_lost++; in rxrpc_rack_mark_lost()
62 static void rxrpc_rack_update(struct rxrpc_call *call, in rxrpc_rack_update() argument
69 ktime_t rtt = ktime_sub(call->acks_latest_ts, xmit_ts); in rxrpc_rack_update()
72 call->tx_nr_lost--; in rxrpc_rack_update()
76 if (before(call->acks_highest_serial, tq->segment_serial[ix])) in rxrpc_rack_update()
78 if (rtt < minmax_get(&call->min_rtt)) in rxrpc_rack_update()
88 if (ktime_after(xmit_ts, call->rack_rtt_ts)) { in rxrpc_rack_update()
89 call->rack_rtt = rtt; in rxrpc_rack_update()
[all …]
Dsendmsg.c23 bool rxrpc_propose_abort(struct rxrpc_call *call, s32 abort_code, int error, in rxrpc_propose_abort() argument
26 _enter("{%d},%d,%d,%u", call->debug_id, abort_code, error, why); in rxrpc_propose_abort()
28 if (!call->send_abort && !rxrpc_call_is_complete(call)) { in rxrpc_propose_abort()
29 call->send_abort_why = why; in rxrpc_propose_abort()
30 call->send_abort_err = error; in rxrpc_propose_abort()
31 call->send_abort_seq = 0; in rxrpc_propose_abort()
32 trace_rxrpc_abort_call(call, abort_code); in rxrpc_propose_abort()
34 smp_store_release(&call->send_abort, abort_code); in rxrpc_propose_abort()
35 rxrpc_poke_call(call, rxrpc_call_poke_abort); in rxrpc_propose_abort()
43 * Wait for a call to become connected. Interruption here doesn't cause the
[all …]
Drxperf.c65 int (*deliver)(struct rxperf_call *call);
74 static int rxperf_deliver_param_block(struct rxperf_call *call);
75 static int rxperf_deliver_request(struct rxperf_call *call);
76 static int rxperf_process_call(struct rxperf_call *call);
82 static inline void rxperf_set_call_state(struct rxperf_call *call, in rxperf_set_call_state() argument
85 call->state = to; in rxperf_set_call_state()
88 static inline void rxperf_set_call_complete(struct rxperf_call *call, in rxperf_set_call_complete() argument
91 if (call->state != RXPERF_CALL_COMPLETE) { in rxperf_set_call_complete()
92 call->abort_code = remote_abort; in rxperf_set_call_complete()
93 call->error = error; in rxperf_set_call_complete()
[all …]
Drecvmsg.c20 * Post a call for attention by the socket or kernel service. Further
23 void rxrpc_notify_socket(struct rxrpc_call *call) in rxrpc_notify_socket() argument
28 _enter("%d", call->debug_id); in rxrpc_notify_socket()
30 if (!list_empty(&call->recvmsg_link)) in rxrpc_notify_socket()
35 rx = rcu_dereference(call->socket); in rxrpc_notify_socket()
38 if (call->notify_rx) { in rxrpc_notify_socket()
39 spin_lock_irq(&call->notify_lock); in rxrpc_notify_socket()
40 call->notify_rx(sk, call, call->user_call_ID); in rxrpc_notify_socket()
41 spin_unlock_irq(&call->notify_lock); in rxrpc_notify_socket()
44 if (list_empty(&call->recvmsg_link)) { in rxrpc_notify_socket()
[all …]
Doutput.c48 static void rxrpc_tx_backoff(struct rxrpc_call *call, int ret) in rxrpc_tx_backoff() argument
51 if (call->tx_backoff < 1000) in rxrpc_tx_backoff()
52 call->tx_backoff += 100; in rxrpc_tx_backoff()
54 call->tx_backoff = 0; in rxrpc_tx_backoff()
60 * lets the far side know we're still interested in this call and helps keep
66 static void rxrpc_set_keepalive(struct rxrpc_call *call, ktime_t now) in rxrpc_set_keepalive() argument
68 ktime_t delay = ms_to_ktime(READ_ONCE(call->next_rx_timo) / 6); in rxrpc_set_keepalive()
70 call->keepalive_at = ktime_add(ktime_get_real(), delay); in rxrpc_set_keepalive()
71 trace_rxrpc_timer_set(call, delay, rxrpc_timer_trace_keepalive); in rxrpc_set_keepalive()
77 static int rxrpc_alloc_ack(struct rxrpc_call *call, size_t sack_size) in rxrpc_alloc_ack() argument
[all …]
Drtt.c18 static u32 rxrpc_rto_min_us(struct rxrpc_call *call) in rxrpc_rto_min_us() argument
23 static u32 __rxrpc_set_rto(const struct rxrpc_call *call) in __rxrpc_set_rto() argument
25 return (call->srtt_us >> 3) + call->rttvar_us; in __rxrpc_set_rto()
43 static void rxrpc_rtt_estimator(struct rxrpc_call *call, long sample_rtt_us) in rxrpc_rtt_estimator() argument
46 u32 srtt = call->srtt_us; in rxrpc_rtt_estimator()
69 m -= (call->mdev_us >> 2); /* similar update on mdev */ in rxrpc_rtt_estimator()
81 m -= (call->mdev_us >> 2); /* similar update on mdev */ in rxrpc_rtt_estimator()
84 call->mdev_us += m; /* mdev = 3/4 mdev + 1/4 new */ in rxrpc_rtt_estimator()
85 if (call->mdev_us > call->mdev_max_us) { in rxrpc_rtt_estimator()
86 call->mdev_max_us = call->mdev_us; in rxrpc_rtt_estimator()
[all …]
Dcall_state.c2 /* Call state changing functions.
11 * Transition a call to the complete state.
13 bool rxrpc_set_call_completion(struct rxrpc_call *call, in rxrpc_set_call_completion() argument
18 if (__rxrpc_call_state(call) == RXRPC_CALL_COMPLETE) in rxrpc_set_call_completion()
21 call->abort_code = abort_code; in rxrpc_set_call_completion()
22 call->error = error; in rxrpc_set_call_completion()
23 call->completion = compl; in rxrpc_set_call_completion()
25 rxrpc_set_call_state(call, RXRPC_CALL_COMPLETE); in rxrpc_set_call_completion()
26 trace_rxrpc_call_complete(call); in rxrpc_set_call_completion()
27 wake_up(&call->waitq); in rxrpc_set_call_completion()
[all …]
Dcall_accept.c2 /* incoming call handling
25 static void rxrpc_dummy_notify(struct sock *sk, struct rxrpc_call *call, in rxrpc_dummy_notify() argument
31 * Preallocate a single service call, connection and peer and, if possible,
41 struct rxrpc_call *call, *xcall; in rxrpc_service_prealloc_one() local
98 call = rxrpc_alloc_call(rx, gfp, debug_id); in rxrpc_service_prealloc_one()
99 if (!call) in rxrpc_service_prealloc_one()
101 call->flags |= (1 << RXRPC_CALL_IS_SERVICE); in rxrpc_service_prealloc_one()
102 rxrpc_set_call_state(call, RXRPC_CALL_SERVER_PREALLOC); in rxrpc_service_prealloc_one()
103 __set_bit(RXRPC_CALL_EV_INITIAL_PING, &call->events); in rxrpc_service_prealloc_one()
105 trace_rxrpc_call(call->debug_id, refcount_read(&call->ref), in rxrpc_service_prealloc_one()
[all …]
Dconn_client.c8 * call so as to handle retransmitted DATA packets in case the server didn't
15 * or a call ID counter overflows.
73 static struct rxrpc_bundle *rxrpc_alloc_bundle(struct rxrpc_call *call, in rxrpc_alloc_bundle() argument
81 bundle->local = call->local; in rxrpc_alloc_bundle()
82 bundle->peer = rxrpc_get_peer(call->peer, rxrpc_peer_get_bundle); in rxrpc_alloc_bundle()
83 bundle->key = key_get(call->key); in rxrpc_alloc_bundle()
84 bundle->security = call->security; in rxrpc_alloc_bundle()
85 bundle->exclusive = test_bit(RXRPC_CALL_EXCLUSIVE, &call->flags); in rxrpc_alloc_bundle()
86 bundle->upgrade = test_bit(RXRPC_CALL_UPGRADE, &call->flags); in rxrpc_alloc_bundle()
87 bundle->service_id = call->dest_srx.srx_service; in rxrpc_alloc_bundle()
[all …]
/linux-6.14.4/include/trace/events/
Drxrpc.h51 EM(rxkad_abort_resp_call_ctr, "rxkad-resp-call-ctr") \
52 EM(rxkad_abort_resp_call_state, "rxkad-resp-call-state") \
72 EM(rxrpc_abort_call_improper_term, "call-improper-term") \
73 EM(rxrpc_abort_call_reset, "call-reset") \
74 EM(rxrpc_abort_call_sendmsg, "call-sendmsg") \
75 EM(rxrpc_abort_call_sock_release, "call-sock-rel") \
76 EM(rxrpc_abort_call_sock_release_tba, "call-sock-rel-tba") \
77 EM(rxrpc_abort_call_timeout, "call-timeout") \
89 EM(rxrpc_badmsg_zero_call, "zero-call") \
101 EM(rxrpc_eproto_no_client_call, "no-cl-call") \
[all …]
/linux-6.14.4/include/trace/
Dtrace_events.h8 * struct trace_event_raw_<call> {
114 * struct trace_event_data_offsets_<call> {
128 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
129 struct trace_event_data_offsets_##call { \
157 * trace_raw_output_<call>(struct trace_iterator *iter, int flags)
160 * struct trace_event_raw_<call> *field; <-- defined in stage 1
169 * if (entry->type != event_<call>->event.type) {
177 * return trace_output_call(iter, <call>, <TP_printk> "\n");
203 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
205 trace_raw_output_##call(struct trace_iterator *iter, int flags, \
[all …]
Dtrace_custom_events.h62 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
63 struct trace_custom_event_data_offsets_##call { \
77 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
79 trace_custom_raw_output_##call(struct trace_iterator *iter, int flags, \
84 struct trace_custom_event_raw_##call *field; \
97 static struct trace_event_functions trace_custom_event_type_funcs_##call = { \
98 .trace = trace_custom_raw_output_##call, \
108 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, func, print) \ argument
109 static struct trace_event_fields trace_custom_event_fields_##call[] = { \
120 #define DECLARE_CUSTOM_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
[all …]
Dbpf_probe.h45 #define __BPF_DECLARE_TRACE(call, proto, args) \ argument
47 __bpf_trace_##call(void *__data, proto) \
53 #define DECLARE_EVENT_CLASS(call, proto, args, tstruct, assign, print) \ argument
54 __BPF_DECLARE_TRACE(call, PARAMS(proto), PARAMS(args))
56 #define __BPF_DECLARE_TRACE_SYSCALL(call, proto, args) \ argument
58 __bpf_trace_##call(void *__data, proto) \
67 #define DECLARE_EVENT_SYSCALL_CLASS(call, proto, args, tstruct, assign, print) \ argument
68 __BPF_DECLARE_TRACE_SYSCALL(call, PARAMS(proto), PARAMS(args))
75 #define __DEFINE_EVENT(template, call, proto, args, size) \ argument
76 static inline void bpf_test_probe_##call(void) \
[all …]
/linux-6.14.4/include/asm-generic/
Dsyscall.h3 * Access to user system call parameters and results
23 * syscall_get_nr - find what system call a task is executing
27 * If @task is executing a system call or is at system call
28 * tracing about to attempt one, returns the system call number.
29 * If @task is not executing a system call, i.e. it's blocked
33 * system call number can be meaningful. If the actual arch value
36 * It's only valid to call this when @task is known to be blocked.
41 * syscall_rollback - roll back registers after an aborted system call
42 * @task: task of interest, must be in system call exit tracing
45 * It's only valid to call this when @task is stopped for system
[all …]
/linux-6.14.4/tools/
DMakefile67 $(call descend,power/$@)
70 $(call descend,power/$@)
73 $(call descend,$@)
76 $(call descend,$@)
79 $(call descend,lib/api)
82 $(call descend,include/nolibc)
85 $(call descend,include/nolibc,$(patsubst nolibc_%,%,$@))
96 $(call descend,sched_ext)
99 $(call descend,testing/$@)
102 $(call descend,lib/$@)
[all …]
/linux-6.14.4/Documentation/networking/
Drxrpc.rst64 (3) Retention of the reusable bits of the transport system set up for one call
122 (#) Each RxRPC operation is a "call". A connection may make up to four
147 explicitly sequenced per call.
158 (#) An call is complete when the request has been sent, the reply has been
162 (#) An call may be aborted by either end at any time up to its completion.
182 the last call currently using it has completed in case a new call is made
215 be used in all other sendmsgs or recvmsgs associated with that call. The
220 first sendmsg() of a call (struct msghdr::msg_name).
226 first sendmsg() of the call must specify the target address. The server's
229 (#) Once the application has received the last message associated with a call,
[all …]
/linux-6.14.4/include/linux/firmware/intel/
Dstratix10-smc.h13 * This file defines the Secure Monitor Call (SMC) message protocol used for
29 * FAST call executes atomic operations, returns when the requested operation
31 * STD call starts a operation which can be preempted by a non-secure
32 * interrupt. The call can return before the requested operation has
51 * Return values in INTEL_SIP_SMC_* call
81 * Sync call used by service driver at EL1 to request the FPGA in EL3 to
84 * Call register usage:
101 * Async call used by service driver at EL1 to provide FPGA configuration data
104 * Call register usage:
127 * Sync call used by service driver at EL1 to track the completed write
[all …]

12345678910>>...226