rxrpc: Permit multiple service binding
[platform/kernel/linux-exynos.git] / net / rxrpc / ar-internal.h
1 /* AF_RXRPC internal definitions
2  *
3  * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4  * Written by David Howells (dhowells@redhat.com)
5  *
6  * This program is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU General Public License
8  * as published by the Free Software Foundation; either version
9  * 2 of the License, or (at your option) any later version.
10  */
11
12 #include <linux/atomic.h>
13 #include <linux/seqlock.h>
14 #include <net/net_namespace.h>
15 #include <net/netns/generic.h>
16 #include <net/sock.h>
17 #include <net/af_rxrpc.h>
18 #include <rxrpc/packet.h>
19
20 #if 0
21 #define CHECK_SLAB_OKAY(X)                                   \
22         BUG_ON(atomic_read((X)) >> (sizeof(atomic_t) - 2) == \
23                (POISON_FREE << 8 | POISON_FREE))
24 #else
25 #define CHECK_SLAB_OKAY(X) do {} while (0)
26 #endif
27
28 #define FCRYPT_BSIZE 8
29 struct rxrpc_crypt {
30         union {
31                 u8      x[FCRYPT_BSIZE];
32                 __be32  n[2];
33         };
34 } __attribute__((aligned(8)));
35
36 #define rxrpc_queue_work(WS)    queue_work(rxrpc_workqueue, (WS))
37 #define rxrpc_queue_delayed_work(WS,D)  \
38         queue_delayed_work(rxrpc_workqueue, (WS), (D))
39
40 struct rxrpc_connection;
41
42 /*
43  * Mark applied to socket buffers.
44  */
45 enum rxrpc_skb_mark {
46         RXRPC_SKB_MARK_DATA,            /* data message */
47         RXRPC_SKB_MARK_FINAL_ACK,       /* final ACK received message */
48         RXRPC_SKB_MARK_BUSY,            /* server busy message */
49         RXRPC_SKB_MARK_REMOTE_ABORT,    /* remote abort message */
50         RXRPC_SKB_MARK_LOCAL_ABORT,     /* local abort message */
51         RXRPC_SKB_MARK_NET_ERROR,       /* network error message */
52         RXRPC_SKB_MARK_LOCAL_ERROR,     /* local error message */
53         RXRPC_SKB_MARK_NEW_CALL,        /* local error message */
54 };
55
56 /*
57  * sk_state for RxRPC sockets
58  */
59 enum {
60         RXRPC_UNBOUND = 0,
61         RXRPC_CLIENT_UNBOUND,           /* Unbound socket used as client */
62         RXRPC_CLIENT_BOUND,             /* client local address bound */
63         RXRPC_SERVER_BOUND,             /* server local address bound */
64         RXRPC_SERVER_BOUND2,            /* second server local address bound */
65         RXRPC_SERVER_LISTENING,         /* server listening for connections */
66         RXRPC_SERVER_LISTEN_DISABLED,   /* server listening disabled */
67         RXRPC_CLOSE,                    /* socket is being closed */
68 };
69
70 /*
71  * Per-network namespace data.
72  */
73 struct rxrpc_net {
74         struct proc_dir_entry   *proc_net;      /* Subdir in /proc/net */
75         u32                     epoch;          /* Local epoch for detecting local-end reset */
76         struct list_head        calls;          /* List of calls active in this namespace */
77         rwlock_t                call_lock;      /* Lock for ->calls */
78
79         struct list_head        conn_proc_list; /* List of conns in this namespace for proc */
80         struct list_head        service_conns;  /* Service conns in this namespace */
81         rwlock_t                conn_lock;      /* Lock for ->conn_proc_list, ->service_conns */
82         struct delayed_work     service_conn_reaper;
83
84         unsigned int            nr_client_conns;
85         unsigned int            nr_active_client_conns;
86         bool                    kill_all_client_conns;
87         spinlock_t              client_conn_cache_lock; /* Lock for ->*_client_conns */
88         spinlock_t              client_conn_discard_lock; /* Prevent multiple discarders */
89         struct list_head        waiting_client_conns;
90         struct list_head        active_client_conns;
91         struct list_head        idle_client_conns;
92         struct delayed_work     client_conn_reaper;
93
94         struct list_head        local_endpoints;
95         struct mutex            local_mutex;    /* Lock for ->local_endpoints */
96
97         spinlock_t              peer_hash_lock; /* Lock for ->peer_hash */
98         DECLARE_HASHTABLE       (peer_hash, 10);
99 };
100
101 /*
102  * Service backlog preallocation.
103  *
104  * This contains circular buffers of preallocated peers, connections and calls
105  * for incoming service calls and their head and tail pointers.  This allows
106  * calls to be set up in the data_ready handler, thereby avoiding the need to
107  * shuffle packets around so much.
108  */
109 struct rxrpc_backlog {
110         unsigned short          peer_backlog_head;
111         unsigned short          peer_backlog_tail;
112         unsigned short          conn_backlog_head;
113         unsigned short          conn_backlog_tail;
114         unsigned short          call_backlog_head;
115         unsigned short          call_backlog_tail;
116 #define RXRPC_BACKLOG_MAX       32
117         struct rxrpc_peer       *peer_backlog[RXRPC_BACKLOG_MAX];
118         struct rxrpc_connection *conn_backlog[RXRPC_BACKLOG_MAX];
119         struct rxrpc_call       *call_backlog[RXRPC_BACKLOG_MAX];
120 };
121
122 /*
123  * RxRPC socket definition
124  */
125 struct rxrpc_sock {
126         /* WARNING: sk has to be the first member */
127         struct sock             sk;
128         rxrpc_notify_new_call_t notify_new_call; /* Func to notify of new call */
129         rxrpc_discard_new_call_t discard_new_call; /* Func to discard a new call */
130         struct rxrpc_local      *local;         /* local endpoint */
131         struct rxrpc_backlog    *backlog;       /* Preallocation for services */
132         spinlock_t              incoming_lock;  /* Incoming call vs service shutdown lock */
133         struct list_head        sock_calls;     /* List of calls owned by this socket */
134         struct list_head        to_be_accepted; /* calls awaiting acceptance */
135         struct list_head        recvmsg_q;      /* Calls awaiting recvmsg's attention  */
136         rwlock_t                recvmsg_lock;   /* Lock for recvmsg_q */
137         struct key              *key;           /* security for this socket */
138         struct key              *securities;    /* list of server security descriptors */
139         struct rb_root          calls;          /* User ID -> call mapping */
140         unsigned long           flags;
141 #define RXRPC_SOCK_CONNECTED            0       /* connect_srx is set */
142         rwlock_t                call_lock;      /* lock for calls */
143         u32                     min_sec_level;  /* minimum security level */
144 #define RXRPC_SECURITY_MAX      RXRPC_SECURITY_ENCRYPT
145         bool                    exclusive;      /* Exclusive connection for a client socket */
146         u16                     second_service; /* Additional service bound to the endpoint */
147         sa_family_t             family;         /* Protocol family created with */
148         struct sockaddr_rxrpc   srx;            /* local address */
149         struct sockaddr_rxrpc   connect_srx;    /* Default client address from connect() */
150 };
151
152 #define rxrpc_sk(__sk) container_of((__sk), struct rxrpc_sock, sk)
153
154 /*
155  * CPU-byteorder normalised Rx packet header.
156  */
157 struct rxrpc_host_header {
158         u32             epoch;          /* client boot timestamp */
159         u32             cid;            /* connection and channel ID */
160         u32             callNumber;     /* call ID (0 for connection-level packets) */
161         u32             seq;            /* sequence number of pkt in call stream */
162         u32             serial;         /* serial number of pkt sent to network */
163         u8              type;           /* packet type */
164         u8              flags;          /* packet flags */
165         u8              userStatus;     /* app-layer defined status */
166         u8              securityIndex;  /* security protocol ID */
167         union {
168                 u16     _rsvd;          /* reserved */
169                 u16     cksum;          /* kerberos security checksum */
170         };
171         u16             serviceId;      /* service ID */
172 } __packed;
173
174 /*
175  * RxRPC socket buffer private variables
176  * - max 48 bytes (struct sk_buff::cb)
177  */
178 struct rxrpc_skb_priv {
179         union {
180                 u8              nr_jumbo;       /* Number of jumbo subpackets */
181         };
182         union {
183                 int             remain;         /* amount of space remaining for next write */
184         };
185
186         struct rxrpc_host_header hdr;           /* RxRPC packet header from this packet */
187 };
188
189 #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb)
190
191 /*
192  * RxRPC security module interface
193  */
194 struct rxrpc_security {
195         const char              *name;          /* name of this service */
196         u8                      security_index; /* security type provided */
197
198         /* Initialise a security service */
199         int (*init)(void);
200
201         /* Clean up a security service */
202         void (*exit)(void);
203
204         /* initialise a connection's security */
205         int (*init_connection_security)(struct rxrpc_connection *);
206
207         /* prime a connection's packet security */
208         int (*prime_packet_security)(struct rxrpc_connection *);
209
210         /* impose security on a packet */
211         int (*secure_packet)(struct rxrpc_call *,
212                              struct sk_buff *,
213                              size_t,
214                              void *);
215
216         /* verify the security on a received packet */
217         int (*verify_packet)(struct rxrpc_call *, struct sk_buff *,
218                              unsigned int, unsigned int, rxrpc_seq_t, u16);
219
220         /* Locate the data in a received packet that has been verified. */
221         void (*locate_data)(struct rxrpc_call *, struct sk_buff *,
222                             unsigned int *, unsigned int *);
223
224         /* issue a challenge */
225         int (*issue_challenge)(struct rxrpc_connection *);
226
227         /* respond to a challenge */
228         int (*respond_to_challenge)(struct rxrpc_connection *,
229                                     struct sk_buff *,
230                                     u32 *);
231
232         /* verify a response */
233         int (*verify_response)(struct rxrpc_connection *,
234                                struct sk_buff *,
235                                u32 *);
236
237         /* clear connection security */
238         void (*clear)(struct rxrpc_connection *);
239 };
240
241 /*
242  * RxRPC local transport endpoint description
243  * - owned by a single AF_RXRPC socket
244  * - pointed to by transport socket struct sk_user_data
245  */
246 struct rxrpc_local {
247         struct rcu_head         rcu;
248         atomic_t                usage;
249         struct rxrpc_net        *rxnet;         /* The network ns in which this resides */
250         struct list_head        link;
251         struct socket           *socket;        /* my UDP socket */
252         struct work_struct      processor;
253         struct rxrpc_sock __rcu *service;       /* Service(s) listening on this endpoint */
254         struct rw_semaphore     defrag_sem;     /* control re-enablement of IP DF bit */
255         struct sk_buff_head     reject_queue;   /* packets awaiting rejection */
256         struct sk_buff_head     event_queue;    /* endpoint event packets awaiting processing */
257         struct rb_root          client_conns;   /* Client connections by socket params */
258         spinlock_t              client_conns_lock; /* Lock for client_conns */
259         spinlock_t              lock;           /* access lock */
260         rwlock_t                services_lock;  /* lock for services list */
261         int                     debug_id;       /* debug ID for printks */
262         bool                    dead;
263         struct sockaddr_rxrpc   srx;            /* local address */
264 };
265
266 /*
267  * RxRPC remote transport endpoint definition
268  * - matched by local endpoint, remote port, address and protocol type
269  */
270 struct rxrpc_peer {
271         struct rcu_head         rcu;            /* This must be first */
272         atomic_t                usage;
273         unsigned long           hash_key;
274         struct hlist_node       hash_link;
275         struct rxrpc_local      *local;
276         struct hlist_head       error_targets;  /* targets for net error distribution */
277         struct work_struct      error_distributor;
278         struct rb_root          service_conns;  /* Service connections */
279         seqlock_t               service_conn_lock;
280         spinlock_t              lock;           /* access lock */
281         unsigned int            if_mtu;         /* interface MTU for this peer */
282         unsigned int            mtu;            /* network MTU for this peer */
283         unsigned int            maxdata;        /* data size (MTU - hdrsize) */
284         unsigned short          hdrsize;        /* header size (IP + UDP + RxRPC) */
285         int                     debug_id;       /* debug ID for printks */
286         int                     error_report;   /* Net (+0) or local (+1000000) to distribute */
287 #define RXRPC_LOCAL_ERROR_OFFSET 1000000
288         struct sockaddr_rxrpc   srx;            /* remote address */
289
290         /* calculated RTT cache */
291 #define RXRPC_RTT_CACHE_SIZE 32
292         ktime_t                 rtt_last_req;   /* Time of last RTT request */
293         u64                     rtt;            /* Current RTT estimate (in nS) */
294         u64                     rtt_sum;        /* Sum of cache contents */
295         u64                     rtt_cache[RXRPC_RTT_CACHE_SIZE]; /* Determined RTT cache */
296         u8                      rtt_cursor;     /* next entry at which to insert */
297         u8                      rtt_usage;      /* amount of cache actually used */
298 };
299
300 /*
301  * Keys for matching a connection.
302  */
303 struct rxrpc_conn_proto {
304         union {
305                 struct {
306                         u32     epoch;          /* epoch of this connection */
307                         u32     cid;            /* connection ID */
308                 };
309                 u64             index_key;
310         };
311 };
312
313 struct rxrpc_conn_parameters {
314         struct rxrpc_local      *local;         /* Representation of local endpoint */
315         struct rxrpc_peer       *peer;          /* Remote endpoint */
316         struct key              *key;           /* Security details */
317         bool                    exclusive;      /* T if conn is exclusive */
318         u16                     service_id;     /* Service ID for this connection */
319         u32                     security_level; /* Security level selected */
320 };
321
322 /*
323  * Bits in the connection flags.
324  */
325 enum rxrpc_conn_flag {
326         RXRPC_CONN_HAS_IDR,             /* Has a client conn ID assigned */
327         RXRPC_CONN_IN_SERVICE_CONNS,    /* Conn is in peer->service_conns */
328         RXRPC_CONN_IN_CLIENT_CONNS,     /* Conn is in local->client_conns */
329         RXRPC_CONN_EXPOSED,             /* Conn has extra ref for exposure */
330         RXRPC_CONN_DONT_REUSE,          /* Don't reuse this connection */
331         RXRPC_CONN_COUNTED,             /* Counted by rxrpc_nr_client_conns */
332 };
333
334 /*
335  * Events that can be raised upon a connection.
336  */
337 enum rxrpc_conn_event {
338         RXRPC_CONN_EV_CHALLENGE,        /* Send challenge packet */
339 };
340
341 /*
342  * The connection cache state.
343  */
344 enum rxrpc_conn_cache_state {
345         RXRPC_CONN_CLIENT_INACTIVE,     /* Conn is not yet listed */
346         RXRPC_CONN_CLIENT_WAITING,      /* Conn is on wait list, waiting for capacity */
347         RXRPC_CONN_CLIENT_ACTIVE,       /* Conn is on active list, doing calls */
348         RXRPC_CONN_CLIENT_CULLED,       /* Conn is culled and delisted, doing calls */
349         RXRPC_CONN_CLIENT_IDLE,         /* Conn is on idle list, doing mostly nothing */
350         RXRPC_CONN__NR_CACHE_STATES
351 };
352
353 /*
354  * The connection protocol state.
355  */
356 enum rxrpc_conn_proto_state {
357         RXRPC_CONN_UNUSED,              /* Connection not yet attempted */
358         RXRPC_CONN_CLIENT,              /* Client connection */
359         RXRPC_CONN_SERVICE_PREALLOC,    /* Service connection preallocation */
360         RXRPC_CONN_SERVICE_UNSECURED,   /* Service unsecured connection */
361         RXRPC_CONN_SERVICE_CHALLENGING, /* Service challenging for security */
362         RXRPC_CONN_SERVICE,             /* Service secured connection */
363         RXRPC_CONN_REMOTELY_ABORTED,    /* Conn aborted by peer */
364         RXRPC_CONN_LOCALLY_ABORTED,     /* Conn aborted locally */
365         RXRPC_CONN__NR_STATES
366 };
367
368 /*
369  * RxRPC connection definition
370  * - matched by { local, peer, epoch, conn_id, direction }
371  * - each connection can only handle four simultaneous calls
372  */
373 struct rxrpc_connection {
374         struct rxrpc_conn_proto proto;
375         struct rxrpc_conn_parameters params;
376
377         atomic_t                usage;
378         struct rcu_head         rcu;
379         struct list_head        cache_link;
380
381         spinlock_t              channel_lock;
382         unsigned char           active_chans;   /* Mask of active channels */
383 #define RXRPC_ACTIVE_CHANS_MASK ((1 << RXRPC_MAXCALLS) - 1)
384         struct list_head        waiting_calls;  /* Calls waiting for channels */
385         struct rxrpc_channel {
386                 struct rxrpc_call __rcu *call;          /* Active call */
387                 u32                     call_id;        /* ID of current call */
388                 u32                     call_counter;   /* Call ID counter */
389                 u32                     last_call;      /* ID of last call */
390                 u8                      last_type;      /* Type of last packet */
391                 union {
392                         u32             last_seq;
393                         u32             last_abort;
394                 };
395         } channels[RXRPC_MAXCALLS];
396
397         struct work_struct      processor;      /* connection event processor */
398         union {
399                 struct rb_node  client_node;    /* Node in local->client_conns */
400                 struct rb_node  service_node;   /* Node in peer->service_conns */
401         };
402         struct list_head        proc_link;      /* link in procfs list */
403         struct list_head        link;           /* link in master connection list */
404         struct sk_buff_head     rx_queue;       /* received conn-level packets */
405         const struct rxrpc_security *security;  /* applied security module */
406         struct key              *server_key;    /* security for this service */
407         struct crypto_skcipher  *cipher;        /* encryption handle */
408         struct rxrpc_crypt      csum_iv;        /* packet checksum base */
409         unsigned long           flags;
410         unsigned long           events;
411         unsigned long           idle_timestamp; /* Time at which last became idle */
412         spinlock_t              state_lock;     /* state-change lock */
413         enum rxrpc_conn_cache_state cache_state;
414         enum rxrpc_conn_proto_state state;      /* current state of connection */
415         u32                     local_abort;    /* local abort code */
416         u32                     remote_abort;   /* remote abort code */
417         int                     debug_id;       /* debug ID for printks */
418         atomic_t                serial;         /* packet serial number counter */
419         unsigned int            hi_serial;      /* highest serial number received */
420         u32                     security_nonce; /* response re-use preventer */
421         u16                     service_id;     /* Service ID, possibly upgraded */
422         u8                      size_align;     /* data size alignment (for security) */
423         u8                      security_size;  /* security header size */
424         u8                      security_ix;    /* security type */
425         u8                      out_clientflag; /* RXRPC_CLIENT_INITIATED if we are client */
426 };
427
428 /*
429  * Flags in call->flags.
430  */
431 enum rxrpc_call_flag {
432         RXRPC_CALL_RELEASED,            /* call has been released - no more message to userspace */
433         RXRPC_CALL_HAS_USERID,          /* has a user ID attached */
434         RXRPC_CALL_IS_SERVICE,          /* Call is service call */
435         RXRPC_CALL_EXPOSED,             /* The call was exposed to the world */
436         RXRPC_CALL_RX_LAST,             /* Received the last packet (at rxtx_top) */
437         RXRPC_CALL_TX_LAST,             /* Last packet in Tx buffer (at rxtx_top) */
438         RXRPC_CALL_SEND_PING,           /* A ping will need to be sent */
439         RXRPC_CALL_PINGING,             /* Ping in process */
440         RXRPC_CALL_RETRANS_TIMEOUT,     /* Retransmission due to timeout occurred */
441 };
442
443 /*
444  * Events that can be raised on a call.
445  */
446 enum rxrpc_call_event {
447         RXRPC_CALL_EV_ACK,              /* need to generate ACK */
448         RXRPC_CALL_EV_ABORT,            /* need to generate abort */
449         RXRPC_CALL_EV_TIMER,            /* Timer expired */
450         RXRPC_CALL_EV_RESEND,           /* Tx resend required */
451         RXRPC_CALL_EV_PING,             /* Ping send required */
452 };
453
454 /*
455  * The states that a call can be in.
456  */
457 enum rxrpc_call_state {
458         RXRPC_CALL_UNINITIALISED,
459         RXRPC_CALL_CLIENT_AWAIT_CONN,   /* - client waiting for connection to become available */
460         RXRPC_CALL_CLIENT_SEND_REQUEST, /* - client sending request phase */
461         RXRPC_CALL_CLIENT_AWAIT_REPLY,  /* - client awaiting reply */
462         RXRPC_CALL_CLIENT_RECV_REPLY,   /* - client receiving reply phase */
463         RXRPC_CALL_SERVER_PREALLOC,     /* - service preallocation */
464         RXRPC_CALL_SERVER_SECURING,     /* - server securing request connection */
465         RXRPC_CALL_SERVER_ACCEPTING,    /* - server accepting request */
466         RXRPC_CALL_SERVER_RECV_REQUEST, /* - server receiving request */
467         RXRPC_CALL_SERVER_ACK_REQUEST,  /* - server pending ACK of request */
468         RXRPC_CALL_SERVER_SEND_REPLY,   /* - server sending reply */
469         RXRPC_CALL_SERVER_AWAIT_ACK,    /* - server awaiting final ACK */
470         RXRPC_CALL_COMPLETE,            /* - call complete */
471         NR__RXRPC_CALL_STATES
472 };
473
474 /*
475  * Call completion condition (state == RXRPC_CALL_COMPLETE).
476  */
477 enum rxrpc_call_completion {
478         RXRPC_CALL_SUCCEEDED,           /* - Normal termination */
479         RXRPC_CALL_REMOTELY_ABORTED,    /* - call aborted by peer */
480         RXRPC_CALL_LOCALLY_ABORTED,     /* - call aborted locally on error or close */
481         RXRPC_CALL_LOCAL_ERROR,         /* - call failed due to local error */
482         RXRPC_CALL_NETWORK_ERROR,       /* - call terminated by network error */
483         NR__RXRPC_CALL_COMPLETIONS
484 };
485
486 /*
487  * Call Tx congestion management modes.
488  */
489 enum rxrpc_congest_mode {
490         RXRPC_CALL_SLOW_START,
491         RXRPC_CALL_CONGEST_AVOIDANCE,
492         RXRPC_CALL_PACKET_LOSS,
493         RXRPC_CALL_FAST_RETRANSMIT,
494         NR__RXRPC_CONGEST_MODES
495 };
496
497 /*
498  * RxRPC call definition
499  * - matched by { connection, call_id }
500  */
501 struct rxrpc_call {
502         struct rcu_head         rcu;
503         struct rxrpc_connection *conn;          /* connection carrying call */
504         struct rxrpc_peer       *peer;          /* Peer record for remote address */
505         struct rxrpc_sock __rcu *socket;        /* socket responsible */
506         struct mutex            user_mutex;     /* User access mutex */
507         ktime_t                 ack_at;         /* When deferred ACK needs to happen */
508         ktime_t                 resend_at;      /* When next resend needs to happen */
509         ktime_t                 ping_at;        /* When next to send a ping */
510         ktime_t                 expire_at;      /* When the call times out */
511         struct timer_list       timer;          /* Combined event timer */
512         struct work_struct      processor;      /* Event processor */
513         rxrpc_notify_rx_t       notify_rx;      /* kernel service Rx notification function */
514         struct list_head        link;           /* link in master call list */
515         struct list_head        chan_wait_link; /* Link in conn->waiting_calls */
516         struct hlist_node       error_link;     /* link in error distribution list */
517         struct list_head        accept_link;    /* Link in rx->acceptq */
518         struct list_head        recvmsg_link;   /* Link in rx->recvmsg_q */
519         struct list_head        sock_link;      /* Link in rx->sock_calls */
520         struct rb_node          sock_node;      /* Node in rx->calls */
521         struct sk_buff          *tx_pending;    /* Tx socket buffer being filled */
522         wait_queue_head_t       waitq;          /* Wait queue for channel or Tx */
523         __be32                  crypto_buf[2];  /* Temporary packet crypto buffer */
524         unsigned long           user_call_ID;   /* user-defined call ID */
525         unsigned long           flags;
526         unsigned long           events;
527         spinlock_t              lock;
528         rwlock_t                state_lock;     /* lock for state transition */
529         u32                     abort_code;     /* Local/remote abort code */
530         int                     error;          /* Local error incurred */
531         enum rxrpc_call_state   state;          /* current state of call */
532         enum rxrpc_call_completion completion;  /* Call completion condition */
533         atomic_t                usage;
534         u16                     service_id;     /* service ID */
535         u8                      security_ix;    /* Security type */
536         u32                     call_id;        /* call ID on connection  */
537         u32                     cid;            /* connection ID plus channel index */
538         int                     debug_id;       /* debug ID for printks */
539         unsigned short          rx_pkt_offset;  /* Current recvmsg packet offset */
540         unsigned short          rx_pkt_len;     /* Current recvmsg packet len */
541
542         /* Rx/Tx circular buffer, depending on phase.
543          *
544          * In the Rx phase, packets are annotated with 0 or the number of the
545          * segment of a jumbo packet each buffer refers to.  There can be up to
546          * 47 segments in a maximum-size UDP packet.
547          *
548          * In the Tx phase, packets are annotated with which buffers have been
549          * acked.
550          */
551 #define RXRPC_RXTX_BUFF_SIZE    64
552 #define RXRPC_RXTX_BUFF_MASK    (RXRPC_RXTX_BUFF_SIZE - 1)
553 #define RXRPC_INIT_RX_WINDOW_SIZE 32
554         struct sk_buff          **rxtx_buffer;
555         u8                      *rxtx_annotations;
556 #define RXRPC_TX_ANNO_ACK       0
557 #define RXRPC_TX_ANNO_UNACK     1
558 #define RXRPC_TX_ANNO_NAK       2
559 #define RXRPC_TX_ANNO_RETRANS   3
560 #define RXRPC_TX_ANNO_MASK      0x03
561 #define RXRPC_TX_ANNO_LAST      0x04
562 #define RXRPC_TX_ANNO_RESENT    0x08
563
564 #define RXRPC_RX_ANNO_JUMBO     0x3f            /* Jumbo subpacket number + 1 if not zero */
565 #define RXRPC_RX_ANNO_JLAST     0x40            /* Set if last element of a jumbo packet */
566 #define RXRPC_RX_ANNO_VERIFIED  0x80            /* Set if verified and decrypted */
567         rxrpc_seq_t             tx_hard_ack;    /* Dead slot in buffer; the first transmitted but
568                                                  * not hard-ACK'd packet follows this.
569                                                  */
570         rxrpc_seq_t             tx_top;         /* Highest Tx slot allocated. */
571
572         /* TCP-style slow-start congestion control [RFC5681].  Since the SMSS
573          * is fixed, we keep these numbers in terms of segments (ie. DATA
574          * packets) rather than bytes.
575          */
576 #define RXRPC_TX_SMSS           RXRPC_JUMBO_DATALEN
577         u8                      cong_cwnd;      /* Congestion window size */
578         u8                      cong_extra;     /* Extra to send for congestion management */
579         u8                      cong_ssthresh;  /* Slow-start threshold */
580         enum rxrpc_congest_mode cong_mode:8;    /* Congestion management mode */
581         u8                      cong_dup_acks;  /* Count of ACKs showing missing packets */
582         u8                      cong_cumul_acks; /* Cumulative ACK count */
583         ktime_t                 cong_tstamp;    /* Last time cwnd was changed */
584
585         rxrpc_seq_t             rx_hard_ack;    /* Dead slot in buffer; the first received but not
586                                                  * consumed packet follows this.
587                                                  */
588         rxrpc_seq_t             rx_top;         /* Highest Rx slot allocated. */
589         rxrpc_seq_t             rx_expect_next; /* Expected next packet sequence number */
590         u8                      rx_winsize;     /* Size of Rx window */
591         u8                      tx_winsize;     /* Maximum size of Tx window */
592         bool                    tx_phase;       /* T if transmission phase, F if receive phase */
593         u8                      nr_jumbo_bad;   /* Number of jumbo dups/exceeds-windows */
594
595         /* receive-phase ACK management */
596         u8                      ackr_reason;    /* reason to ACK */
597         u16                     ackr_skew;      /* skew on packet being ACK'd */
598         rxrpc_serial_t          ackr_serial;    /* serial of packet being ACK'd */
599         rxrpc_seq_t             ackr_prev_seq;  /* previous sequence number received */
600         rxrpc_seq_t             ackr_consumed;  /* Highest packet shown consumed */
601         rxrpc_seq_t             ackr_seen;      /* Highest packet shown seen */
602
603         /* ping management */
604         rxrpc_serial_t          ping_serial;    /* Last ping sent */
605         ktime_t                 ping_time;      /* Time last ping sent */
606
607         /* transmission-phase ACK management */
608         ktime_t                 acks_latest_ts; /* Timestamp of latest ACK received */
609         rxrpc_serial_t          acks_latest;    /* serial number of latest ACK received */
610         rxrpc_seq_t             acks_lowest_nak; /* Lowest NACK in the buffer (or ==tx_hard_ack) */
611 };
612
613 /*
614  * Summary of a new ACK and the changes it made to the Tx buffer packet states.
615  */
616 struct rxrpc_ack_summary {
617         u8                      ack_reason;
618         u8                      nr_acks;                /* Number of ACKs in packet */
619         u8                      nr_nacks;               /* Number of NACKs in packet */
620         u8                      nr_new_acks;            /* Number of new ACKs in packet */
621         u8                      nr_new_nacks;           /* Number of new NACKs in packet */
622         u8                      nr_rot_new_acks;        /* Number of rotated new ACKs */
623         bool                    new_low_nack;           /* T if new low NACK found */
624         bool                    retrans_timeo;          /* T if reTx due to timeout happened */
625         u8                      flight_size;            /* Number of unreceived transmissions */
626         /* Place to stash values for tracing */
627         enum rxrpc_congest_mode mode:8;
628         u8                      cwnd;
629         u8                      ssthresh;
630         u8                      dup_acks;
631         u8                      cumulative_acks;
632 };
633
634 #include <trace/events/rxrpc.h>
635
636 /*
637  * af_rxrpc.c
638  */
639 extern atomic_t rxrpc_n_tx_skbs, rxrpc_n_rx_skbs;
640 extern atomic_t rxrpc_debug_id;
641 extern struct workqueue_struct *rxrpc_workqueue;
642
643 /*
644  * call_accept.c
645  */
646 int rxrpc_service_prealloc(struct rxrpc_sock *, gfp_t);
647 void rxrpc_discard_prealloc(struct rxrpc_sock *);
648 struct rxrpc_call *rxrpc_new_incoming_call(struct rxrpc_local *,
649                                            struct rxrpc_connection *,
650                                            struct sk_buff *);
651 void rxrpc_accept_incoming_calls(struct rxrpc_local *);
652 struct rxrpc_call *rxrpc_accept_call(struct rxrpc_sock *, unsigned long,
653                                      rxrpc_notify_rx_t);
654 int rxrpc_reject_call(struct rxrpc_sock *);
655
656 /*
657  * call_event.c
658  */
659 void __rxrpc_set_timer(struct rxrpc_call *, enum rxrpc_timer_trace, ktime_t);
660 void rxrpc_set_timer(struct rxrpc_call *, enum rxrpc_timer_trace, ktime_t);
661 void rxrpc_propose_ACK(struct rxrpc_call *, u8, u16, u32, bool, bool,
662                        enum rxrpc_propose_ack_trace);
663 void rxrpc_process_call(struct work_struct *);
664
665 /*
666  * call_object.c
667  */
668 extern const char *const rxrpc_call_states[];
669 extern const char *const rxrpc_call_completions[];
670 extern unsigned int rxrpc_max_call_lifetime;
671 extern struct kmem_cache *rxrpc_call_jar;
672
673 struct rxrpc_call *rxrpc_find_call_by_user_ID(struct rxrpc_sock *, unsigned long);
674 struct rxrpc_call *rxrpc_alloc_call(gfp_t);
675 struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *,
676                                          struct rxrpc_conn_parameters *,
677                                          struct sockaddr_rxrpc *,
678                                          unsigned long, gfp_t);
679 void rxrpc_incoming_call(struct rxrpc_sock *, struct rxrpc_call *,
680                          struct sk_buff *);
681 void rxrpc_release_call(struct rxrpc_sock *, struct rxrpc_call *);
682 void rxrpc_release_calls_on_socket(struct rxrpc_sock *);
683 bool __rxrpc_queue_call(struct rxrpc_call *);
684 bool rxrpc_queue_call(struct rxrpc_call *);
685 void rxrpc_see_call(struct rxrpc_call *);
686 void rxrpc_get_call(struct rxrpc_call *, enum rxrpc_call_trace);
687 void rxrpc_put_call(struct rxrpc_call *, enum rxrpc_call_trace);
688 void rxrpc_cleanup_call(struct rxrpc_call *);
689 void rxrpc_destroy_all_calls(struct rxrpc_net *);
690
691 static inline bool rxrpc_is_service_call(const struct rxrpc_call *call)
692 {
693         return test_bit(RXRPC_CALL_IS_SERVICE, &call->flags);
694 }
695
696 static inline bool rxrpc_is_client_call(const struct rxrpc_call *call)
697 {
698         return !rxrpc_is_service_call(call);
699 }
700
701 /*
702  * Transition a call to the complete state.
703  */
704 static inline bool __rxrpc_set_call_completion(struct rxrpc_call *call,
705                                                enum rxrpc_call_completion compl,
706                                                u32 abort_code,
707                                                int error)
708 {
709         if (call->state < RXRPC_CALL_COMPLETE) {
710                 call->abort_code = abort_code;
711                 call->error = error;
712                 call->completion = compl,
713                 call->state = RXRPC_CALL_COMPLETE;
714                 wake_up(&call->waitq);
715                 return true;
716         }
717         return false;
718 }
719
720 static inline bool rxrpc_set_call_completion(struct rxrpc_call *call,
721                                              enum rxrpc_call_completion compl,
722                                              u32 abort_code,
723                                              int error)
724 {
725         bool ret;
726
727         write_lock_bh(&call->state_lock);
728         ret = __rxrpc_set_call_completion(call, compl, abort_code, error);
729         write_unlock_bh(&call->state_lock);
730         return ret;
731 }
732
733 /*
734  * Record that a call successfully completed.
735  */
736 static inline bool __rxrpc_call_completed(struct rxrpc_call *call)
737 {
738         return __rxrpc_set_call_completion(call, RXRPC_CALL_SUCCEEDED, 0, 0);
739 }
740
741 static inline bool rxrpc_call_completed(struct rxrpc_call *call)
742 {
743         bool ret;
744
745         write_lock_bh(&call->state_lock);
746         ret = __rxrpc_call_completed(call);
747         write_unlock_bh(&call->state_lock);
748         return ret;
749 }
750
751 /*
752  * Record that a call is locally aborted.
753  */
754 static inline bool __rxrpc_abort_call(const char *why, struct rxrpc_call *call,
755                                       rxrpc_seq_t seq,
756                                       u32 abort_code, int error)
757 {
758         trace_rxrpc_abort(why, call->cid, call->call_id, seq,
759                           abort_code, error);
760         return __rxrpc_set_call_completion(call, RXRPC_CALL_LOCALLY_ABORTED,
761                                            abort_code, error);
762 }
763
764 static inline bool rxrpc_abort_call(const char *why, struct rxrpc_call *call,
765                                     rxrpc_seq_t seq, u32 abort_code, int error)
766 {
767         bool ret;
768
769         write_lock_bh(&call->state_lock);
770         ret = __rxrpc_abort_call(why, call, seq, abort_code, error);
771         write_unlock_bh(&call->state_lock);
772         return ret;
773 }
774
775 /*
776  * Abort a call due to a protocol error.
777  */
778 static inline bool __rxrpc_abort_eproto(struct rxrpc_call *call,
779                                         struct sk_buff *skb,
780                                         const char *eproto_why,
781                                         const char *why,
782                                         u32 abort_code)
783 {
784         struct rxrpc_skb_priv *sp = rxrpc_skb(skb);
785
786         trace_rxrpc_rx_eproto(call, sp->hdr.serial, eproto_why);
787         return rxrpc_abort_call(why, call, sp->hdr.seq, abort_code, -EPROTO);
788 }
789
790 #define rxrpc_abort_eproto(call, skb, eproto_why, abort_why, abort_code) \
791         __rxrpc_abort_eproto((call), (skb), tracepoint_string(eproto_why), \
792                              (abort_why), (abort_code))
793
794 /*
795  * conn_client.c
796  */
797 extern unsigned int rxrpc_max_client_connections;
798 extern unsigned int rxrpc_reap_client_connections;
799 extern unsigned int rxrpc_conn_idle_client_expiry;
800 extern unsigned int rxrpc_conn_idle_client_fast_expiry;
801 extern struct idr rxrpc_client_conn_ids;
802
803 void rxrpc_destroy_client_conn_ids(void);
804 int rxrpc_connect_call(struct rxrpc_call *, struct rxrpc_conn_parameters *,
805                        struct sockaddr_rxrpc *, gfp_t);
806 void rxrpc_expose_client_call(struct rxrpc_call *);
807 void rxrpc_disconnect_client_call(struct rxrpc_call *);
808 void rxrpc_put_client_conn(struct rxrpc_connection *);
809 void rxrpc_discard_expired_client_conns(struct work_struct *);
810 void rxrpc_destroy_all_client_connections(struct rxrpc_net *);
811
812 /*
813  * conn_event.c
814  */
815 void rxrpc_process_connection(struct work_struct *);
816
817 /*
818  * conn_object.c
819  */
820 extern unsigned int rxrpc_connection_expiry;
821
822 int rxrpc_extract_addr_from_skb(struct sockaddr_rxrpc *, struct sk_buff *);
823 struct rxrpc_connection *rxrpc_alloc_connection(gfp_t);
824 struct rxrpc_connection *rxrpc_find_connection_rcu(struct rxrpc_local *,
825                                                    struct sk_buff *);
826 void __rxrpc_disconnect_call(struct rxrpc_connection *, struct rxrpc_call *);
827 void rxrpc_disconnect_call(struct rxrpc_call *);
828 void rxrpc_kill_connection(struct rxrpc_connection *);
829 bool rxrpc_queue_conn(struct rxrpc_connection *);
830 void rxrpc_see_connection(struct rxrpc_connection *);
831 void rxrpc_get_connection(struct rxrpc_connection *);
832 struct rxrpc_connection *rxrpc_get_connection_maybe(struct rxrpc_connection *);
833 void rxrpc_put_service_conn(struct rxrpc_connection *);
834 void rxrpc_service_connection_reaper(struct work_struct *);
835 void rxrpc_destroy_all_connections(struct rxrpc_net *);
836
837 static inline bool rxrpc_conn_is_client(const struct rxrpc_connection *conn)
838 {
839         return conn->out_clientflag;
840 }
841
842 static inline bool rxrpc_conn_is_service(const struct rxrpc_connection *conn)
843 {
844         return !rxrpc_conn_is_client(conn);
845 }
846
847 static inline void rxrpc_put_connection(struct rxrpc_connection *conn)
848 {
849         if (!conn)
850                 return;
851
852         if (rxrpc_conn_is_client(conn))
853                 rxrpc_put_client_conn(conn);
854         else
855                 rxrpc_put_service_conn(conn);
856 }
857
858 /*
859  * conn_service.c
860  */
861 struct rxrpc_connection *rxrpc_find_service_conn_rcu(struct rxrpc_peer *,
862                                                      struct sk_buff *);
863 struct rxrpc_connection *rxrpc_prealloc_service_connection(struct rxrpc_net *, gfp_t);
864 void rxrpc_new_incoming_connection(struct rxrpc_connection *, struct sk_buff *);
865 void rxrpc_unpublish_service_conn(struct rxrpc_connection *);
866
867 /*
868  * input.c
869  */
870 void rxrpc_data_ready(struct sock *);
871
872 /*
873  * insecure.c
874  */
875 extern const struct rxrpc_security rxrpc_no_security;
876
877 /*
878  * key.c
879  */
880 extern struct key_type key_type_rxrpc;
881 extern struct key_type key_type_rxrpc_s;
882
883 int rxrpc_request_key(struct rxrpc_sock *, char __user *, int);
884 int rxrpc_server_keyring(struct rxrpc_sock *, char __user *, int);
885 int rxrpc_get_server_data_key(struct rxrpc_connection *, const void *, time_t,
886                               u32);
887
888 /*
889  * local_event.c
890  */
891 extern void rxrpc_process_local_events(struct rxrpc_local *);
892
893 /*
894  * local_object.c
895  */
896 struct rxrpc_local *rxrpc_lookup_local(struct net *, const struct sockaddr_rxrpc *);
897 void __rxrpc_put_local(struct rxrpc_local *);
898 void rxrpc_destroy_all_locals(struct rxrpc_net *);
899
900 static inline void rxrpc_get_local(struct rxrpc_local *local)
901 {
902         atomic_inc(&local->usage);
903 }
904
905 static inline
906 struct rxrpc_local *rxrpc_get_local_maybe(struct rxrpc_local *local)
907 {
908         return atomic_inc_not_zero(&local->usage) ? local : NULL;
909 }
910
911 static inline void rxrpc_put_local(struct rxrpc_local *local)
912 {
913         if (local && atomic_dec_and_test(&local->usage))
914                 __rxrpc_put_local(local);
915 }
916
917 static inline void rxrpc_queue_local(struct rxrpc_local *local)
918 {
919         rxrpc_queue_work(&local->processor);
920 }
921
922 /*
923  * misc.c
924  */
925 extern unsigned int rxrpc_max_backlog __read_mostly;
926 extern unsigned int rxrpc_requested_ack_delay;
927 extern unsigned int rxrpc_soft_ack_delay;
928 extern unsigned int rxrpc_idle_ack_delay;
929 extern unsigned int rxrpc_rx_window_size;
930 extern unsigned int rxrpc_rx_mtu;
931 extern unsigned int rxrpc_rx_jumbo_max;
932 extern unsigned int rxrpc_resend_timeout;
933
934 extern const s8 rxrpc_ack_priority[];
935
936 /*
937  * net_ns.c
938  */
939 extern unsigned int rxrpc_net_id;
940 extern struct pernet_operations rxrpc_net_ops;
941
942 static inline struct rxrpc_net *rxrpc_net(struct net *net)
943 {
944         return net_generic(net, rxrpc_net_id);
945 }
946
947 /*
948  * output.c
949  */
950 int rxrpc_send_ack_packet(struct rxrpc_call *, bool);
951 int rxrpc_send_abort_packet(struct rxrpc_call *);
952 int rxrpc_send_data_packet(struct rxrpc_call *, struct sk_buff *, bool);
953 void rxrpc_reject_packets(struct rxrpc_local *);
954
955 /*
956  * peer_event.c
957  */
958 void rxrpc_error_report(struct sock *);
959 void rxrpc_peer_error_distributor(struct work_struct *);
960 void rxrpc_peer_add_rtt(struct rxrpc_call *, enum rxrpc_rtt_rx_trace,
961                         rxrpc_serial_t, rxrpc_serial_t, ktime_t, ktime_t);
962
963 /*
964  * peer_object.c
965  */
966 struct rxrpc_peer *rxrpc_lookup_peer_rcu(struct rxrpc_local *,
967                                          const struct sockaddr_rxrpc *);
968 struct rxrpc_peer *rxrpc_lookup_peer(struct rxrpc_local *,
969                                      struct sockaddr_rxrpc *, gfp_t);
970 struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *, gfp_t);
971 struct rxrpc_peer *rxrpc_lookup_incoming_peer(struct rxrpc_local *,
972                                               struct rxrpc_peer *);
973
974 static inline struct rxrpc_peer *rxrpc_get_peer(struct rxrpc_peer *peer)
975 {
976         atomic_inc(&peer->usage);
977         return peer;
978 }
979
980 static inline
981 struct rxrpc_peer *rxrpc_get_peer_maybe(struct rxrpc_peer *peer)
982 {
983         return atomic_inc_not_zero(&peer->usage) ? peer : NULL;
984 }
985
986 extern void __rxrpc_put_peer(struct rxrpc_peer *peer);
987 static inline void rxrpc_put_peer(struct rxrpc_peer *peer)
988 {
989         if (peer && atomic_dec_and_test(&peer->usage))
990                 __rxrpc_put_peer(peer);
991 }
992
993 /*
994  * proc.c
995  */
996 extern const struct file_operations rxrpc_call_seq_fops;
997 extern const struct file_operations rxrpc_connection_seq_fops;
998
999 /*
1000  * recvmsg.c
1001  */
1002 void rxrpc_notify_socket(struct rxrpc_call *);
1003 int rxrpc_recvmsg(struct socket *, struct msghdr *, size_t, int);
1004
1005 /*
1006  * rxkad.c
1007  */
1008 #ifdef CONFIG_RXKAD
1009 extern const struct rxrpc_security rxkad;
1010 #endif
1011
1012 /*
1013  * security.c
1014  */
1015 int __init rxrpc_init_security(void);
1016 void rxrpc_exit_security(void);
1017 int rxrpc_init_client_conn_security(struct rxrpc_connection *);
1018 int rxrpc_init_server_conn_security(struct rxrpc_connection *);
1019
1020 /*
1021  * sendmsg.c
1022  */
1023 int rxrpc_do_sendmsg(struct rxrpc_sock *, struct msghdr *, size_t);
1024
1025 /*
1026  * skbuff.c
1027  */
1028 void rxrpc_kernel_data_consumed(struct rxrpc_call *, struct sk_buff *);
1029 void rxrpc_packet_destructor(struct sk_buff *);
1030 void rxrpc_new_skb(struct sk_buff *, enum rxrpc_skb_trace);
1031 void rxrpc_see_skb(struct sk_buff *, enum rxrpc_skb_trace);
1032 void rxrpc_get_skb(struct sk_buff *, enum rxrpc_skb_trace);
1033 void rxrpc_free_skb(struct sk_buff *, enum rxrpc_skb_trace);
1034 void rxrpc_lose_skb(struct sk_buff *, enum rxrpc_skb_trace);
1035 void rxrpc_purge_queue(struct sk_buff_head *);
1036
1037 /*
1038  * sysctl.c
1039  */
1040 #ifdef CONFIG_SYSCTL
1041 extern int __init rxrpc_sysctl_init(void);
1042 extern void rxrpc_sysctl_exit(void);
1043 #else
1044 static inline int __init rxrpc_sysctl_init(void) { return 0; }
1045 static inline void rxrpc_sysctl_exit(void) {}
1046 #endif
1047
1048 /*
1049  * utils.c
1050  */
1051 int rxrpc_extract_addr_from_skb(struct sockaddr_rxrpc *, struct sk_buff *);
1052
1053 static inline bool before(u32 seq1, u32 seq2)
1054 {
1055         return (s32)(seq1 - seq2) < 0;
1056 }
1057 static inline bool before_eq(u32 seq1, u32 seq2)
1058 {
1059         return (s32)(seq1 - seq2) <= 0;
1060 }
1061 static inline bool after(u32 seq1, u32 seq2)
1062 {
1063         return (s32)(seq1 - seq2) > 0;
1064 }
1065 static inline bool after_eq(u32 seq1, u32 seq2)
1066 {
1067         return (s32)(seq1 - seq2) >= 0;
1068 }
1069
1070 /*
1071  * debug tracing
1072  */
1073 extern unsigned int rxrpc_debug;
1074
1075 #define dbgprintk(FMT,...) \
1076         printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__)
1077
1078 #define kenter(FMT,...) dbgprintk("==> %s("FMT")",__func__ ,##__VA_ARGS__)
1079 #define kleave(FMT,...) dbgprintk("<== %s()"FMT"",__func__ ,##__VA_ARGS__)
1080 #define kdebug(FMT,...) dbgprintk("    "FMT ,##__VA_ARGS__)
1081 #define kproto(FMT,...) dbgprintk("### "FMT ,##__VA_ARGS__)
1082 #define knet(FMT,...)   dbgprintk("@@@ "FMT ,##__VA_ARGS__)
1083
1084
1085 #if defined(__KDEBUG)
1086 #define _enter(FMT,...) kenter(FMT,##__VA_ARGS__)
1087 #define _leave(FMT,...) kleave(FMT,##__VA_ARGS__)
1088 #define _debug(FMT,...) kdebug(FMT,##__VA_ARGS__)
1089 #define _proto(FMT,...) kproto(FMT,##__VA_ARGS__)
1090 #define _net(FMT,...)   knet(FMT,##__VA_ARGS__)
1091
1092 #elif defined(CONFIG_AF_RXRPC_DEBUG)
1093 #define RXRPC_DEBUG_KENTER      0x01
1094 #define RXRPC_DEBUG_KLEAVE      0x02
1095 #define RXRPC_DEBUG_KDEBUG      0x04
1096 #define RXRPC_DEBUG_KPROTO      0x08
1097 #define RXRPC_DEBUG_KNET        0x10
1098
1099 #define _enter(FMT,...)                                 \
1100 do {                                                    \
1101         if (unlikely(rxrpc_debug & RXRPC_DEBUG_KENTER)) \
1102                 kenter(FMT,##__VA_ARGS__);              \
1103 } while (0)
1104
1105 #define _leave(FMT,...)                                 \
1106 do {                                                    \
1107         if (unlikely(rxrpc_debug & RXRPC_DEBUG_KLEAVE)) \
1108                 kleave(FMT,##__VA_ARGS__);              \
1109 } while (0)
1110
1111 #define _debug(FMT,...)                                 \
1112 do {                                                    \
1113         if (unlikely(rxrpc_debug & RXRPC_DEBUG_KDEBUG)) \
1114                 kdebug(FMT,##__VA_ARGS__);              \
1115 } while (0)
1116
1117 #define _proto(FMT,...)                                 \
1118 do {                                                    \
1119         if (unlikely(rxrpc_debug & RXRPC_DEBUG_KPROTO)) \
1120                 kproto(FMT,##__VA_ARGS__);              \
1121 } while (0)
1122
1123 #define _net(FMT,...)                                   \
1124 do {                                                    \
1125         if (unlikely(rxrpc_debug & RXRPC_DEBUG_KNET))   \
1126                 knet(FMT,##__VA_ARGS__);                \
1127 } while (0)
1128
1129 #else
1130 #define _enter(FMT,...) no_printk("==> %s("FMT")",__func__ ,##__VA_ARGS__)
1131 #define _leave(FMT,...) no_printk("<== %s()"FMT"",__func__ ,##__VA_ARGS__)
1132 #define _debug(FMT,...) no_printk("    "FMT ,##__VA_ARGS__)
1133 #define _proto(FMT,...) no_printk("### "FMT ,##__VA_ARGS__)
1134 #define _net(FMT,...)   no_printk("@@@ "FMT ,##__VA_ARGS__)
1135 #endif
1136
1137 /*
1138  * debug assertion checking
1139  */
1140 #if 1 // defined(__KDEBUGALL)
1141
1142 #define ASSERT(X)                                               \
1143 do {                                                            \
1144         if (unlikely(!(X))) {                                   \
1145                 pr_err("Assertion failed\n");                   \
1146                 BUG();                                          \
1147         }                                                       \
1148 } while (0)
1149
1150 #define ASSERTCMP(X, OP, Y)                                             \
1151 do {                                                                    \
1152         __typeof__(X) _x = (X);                                         \
1153         __typeof__(Y) _y = (__typeof__(X))(Y);                          \
1154         if (unlikely(!(_x OP _y))) {                                    \
1155                 pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \
1156                        (unsigned long)_x, (unsigned long)_x, #OP,       \
1157                        (unsigned long)_y, (unsigned long)_y);           \
1158                 BUG();                                                  \
1159         }                                                               \
1160 } while (0)
1161
1162 #define ASSERTIF(C, X)                                          \
1163 do {                                                            \
1164         if (unlikely((C) && !(X))) {                            \
1165                 pr_err("Assertion failed\n");                   \
1166                 BUG();                                          \
1167         }                                                       \
1168 } while (0)
1169
1170 #define ASSERTIFCMP(C, X, OP, Y)                                        \
1171 do {                                                                    \
1172         __typeof__(X) _x = (X);                                         \
1173         __typeof__(Y) _y = (__typeof__(X))(Y);                          \
1174         if (unlikely((C) && !(_x OP _y))) {                             \
1175                 pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \
1176                        (unsigned long)_x, (unsigned long)_x, #OP,       \
1177                        (unsigned long)_y, (unsigned long)_y);           \
1178                 BUG();                                                  \
1179         }                                                               \
1180 } while (0)
1181
1182 #else
1183
1184 #define ASSERT(X)                               \
1185 do {                                            \
1186 } while (0)
1187
1188 #define ASSERTCMP(X, OP, Y)                     \
1189 do {                                            \
1190 } while (0)
1191
1192 #define ASSERTIF(C, X)                          \
1193 do {                                            \
1194 } while (0)
1195
1196 #define ASSERTIFCMP(C, X, OP, Y)                \
1197 do {                                            \
1198 } while (0)
1199
1200 #endif /* __KDEBUGALL */