1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 *
4 * Copyright (C) Alan Cox GW4PTS (alan@lxorguk.ukuu.org.uk)
5 * Copyright (C) Jonathan Naylor G4KLX (g4klx@g4klx.demon.co.uk)
6 * Copyright (C) Joerg Reuter DL1BKE (jreuter@yaina.de)
7 * Copyright (C) Hans-Joachim Hetscher DD8NE (dd8ne@bnv-bamberg.de)
8 */
9 #include <linux/errno.h>
10 #include <linux/types.h>
11 #include <linux/socket.h>
12 #include <linux/in.h>
13 #include <linux/kernel.h>
14 #include <linux/timer.h>
15 #include <linux/string.h>
16 #include <linux/sockios.h>
17 #include <linux/net.h>
18 #include <linux/slab.h>
19 #include <net/ax25.h>
20 #include <linux/inet.h>
21 #include <linux/netdevice.h>
22 #include <linux/skbuff.h>
23 #include <net/sock.h>
24 #include <net/tcp_states.h>
25 #include <linux/uaccess.h>
26 #include <linux/fcntl.h>
27 #include <linux/mm.h>
28 #include <linux/interrupt.h>
29
30 /*
31 * Given a fragment, queue it on the fragment queue and if the fragment
32 * is complete, send it back to ax25_rx_iframe.
33 */
ax25_rx_fragment(ax25_cb * ax25,struct sk_buff * skb)34 static int ax25_rx_fragment(ax25_cb *ax25, struct sk_buff *skb)
35 {
36 struct sk_buff *skbn, *skbo;
37
38 if (ax25->fragno != 0) {
39 if (!(*skb->data & AX25_SEG_FIRST)) {
40 if ((ax25->fragno - 1) == (*skb->data & AX25_SEG_REM)) {
41 /* Enqueue fragment */
42 ax25->fragno = *skb->data & AX25_SEG_REM;
43 skb_pull(skb, 1); /* skip fragno */
44 ax25->fraglen += skb->len;
45 skb_queue_tail(&ax25->frag_queue, skb);
46
47 /* Last fragment received ? */
48 if (ax25->fragno == 0) {
49 skbn = alloc_skb(AX25_MAX_HEADER_LEN +
50 ax25->fraglen,
51 GFP_ATOMIC);
52 if (!skbn) {
53 skb_queue_purge(&ax25->frag_queue);
54 return 1;
55 }
56
57 skb_reserve(skbn, AX25_MAX_HEADER_LEN);
58
59 skbn->dev = ax25->ax25_dev->dev;
60 skb_reset_network_header(skbn);
61 skb_reset_transport_header(skbn);
62
63 /* Copy data from the fragments */
64 while ((skbo = skb_dequeue(&ax25->frag_queue)) != NULL) {
65 skb_copy_from_linear_data(skbo,
66 skb_put(skbn, skbo->len),
67 skbo->len);
68 kfree_skb(skbo);
69 }
70
71 ax25->fraglen = 0;
72
73 if (ax25_rx_iframe(ax25, skbn) == 0)
74 kfree_skb(skbn);
75 }
76
77 return 1;
78 }
79 }
80 } else {
81 /* First fragment received */
82 if (*skb->data & AX25_SEG_FIRST) {
83 skb_queue_purge(&ax25->frag_queue);
84 ax25->fragno = *skb->data & AX25_SEG_REM;
85 skb_pull(skb, 1); /* skip fragno */
86 ax25->fraglen = skb->len;
87 skb_queue_tail(&ax25->frag_queue, skb);
88 return 1;
89 }
90 }
91
92 return 0;
93 }
94
95 /*
96 * This is where all valid I frames are sent to, to be dispatched to
97 * whichever protocol requires them.
98 */
ax25_rx_iframe(ax25_cb * ax25,struct sk_buff * skb)99 int ax25_rx_iframe(ax25_cb *ax25, struct sk_buff *skb)
100 {
101 int (*func)(struct sk_buff *, ax25_cb *);
102 unsigned char pid;
103 int queued = 0;
104
105 if (skb == NULL) return 0;
106
107 ax25_start_idletimer(ax25);
108
109 pid = *skb->data;
110
111 if (pid == AX25_P_IP) {
112 /* working around a TCP bug to keep additional listeners
113 * happy. TCP re-uses the buffer and destroys the original
114 * content.
115 */
116 struct sk_buff *skbn = skb_copy(skb, GFP_ATOMIC);
117 if (skbn != NULL) {
118 kfree_skb(skb);
119 skb = skbn;
120 }
121
122 skb_pull(skb, 1); /* Remove PID */
123 skb->mac_header = skb->network_header;
124 skb_reset_network_header(skb);
125 skb->dev = ax25->ax25_dev->dev;
126 skb->pkt_type = PACKET_HOST;
127 skb->protocol = htons(ETH_P_IP);
128 netif_rx(skb);
129 return 1;
130 }
131 if (pid == AX25_P_SEGMENT) {
132 skb_pull(skb, 1); /* Remove PID */
133 return ax25_rx_fragment(ax25, skb);
134 }
135
136 if ((func = ax25_protocol_function(pid)) != NULL) {
137 skb_pull(skb, 1); /* Remove PID */
138 return (*func)(skb, ax25);
139 }
140
141 if (ax25->sk != NULL && ax25->ax25_dev->values[AX25_VALUES_CONMODE] == 2) {
142 if ((!ax25->pidincl && ax25->sk->sk_protocol == pid) ||
143 ax25->pidincl) {
144 if (sock_queue_rcv_skb(ax25->sk, skb) == 0)
145 queued = 1;
146 else
147 ax25->condition |= AX25_COND_OWN_RX_BUSY;
148 }
149 }
150
151 return queued;
152 }
153
154 /*
155 * Higher level upcall for a LAPB frame
156 */
ax25_process_rx_frame(ax25_cb * ax25,struct sk_buff * skb,int type,int dama)157 static int ax25_process_rx_frame(ax25_cb *ax25, struct sk_buff *skb, int type, int dama)
158 {
159 int queued = 0;
160
161 if (ax25->state == AX25_STATE_0)
162 return 0;
163
164 switch (ax25->ax25_dev->values[AX25_VALUES_PROTOCOL]) {
165 case AX25_PROTO_STD_SIMPLEX:
166 case AX25_PROTO_STD_DUPLEX:
167 queued = ax25_std_frame_in(ax25, skb, type);
168 break;
169
170 #ifdef CONFIG_AX25_DAMA_SLAVE
171 case AX25_PROTO_DAMA_SLAVE:
172 if (dama || ax25->ax25_dev->dama.slave)
173 queued = ax25_ds_frame_in(ax25, skb, type);
174 else
175 queued = ax25_std_frame_in(ax25, skb, type);
176 break;
177 #endif
178 }
179
180 return queued;
181 }
182
ax25_rcv(struct sk_buff * skb,struct net_device * dev,const ax25_address * dev_addr,struct packet_type * ptype)183 static int ax25_rcv(struct sk_buff *skb, struct net_device *dev,
184 const ax25_address *dev_addr, struct packet_type *ptype)
185 {
186 ax25_address src, dest, *next_digi = NULL;
187 int type = 0, mine = 0, dama;
188 struct sock *make, *sk;
189 ax25_digi dp, reverse_dp;
190 ax25_cb *ax25;
191 ax25_dev *ax25_dev;
192
193 /*
194 * Process the AX.25/LAPB frame.
195 */
196
197 skb_reset_transport_header(skb);
198
199 if ((ax25_dev = ax25_dev_ax25dev(dev)) == NULL)
200 goto free;
201
202 /*
203 * Parse the address header.
204 */
205
206 if (ax25_addr_parse(skb->data, skb->len, &src, &dest, &dp, &type, &dama) == NULL)
207 goto free;
208
209 /*
210 * Ours perhaps ?
211 */
212 if (dp.lastrepeat + 1 < dp.ndigi) /* Not yet digipeated completely */
213 next_digi = &dp.calls[dp.lastrepeat + 1];
214
215 /*
216 * Pull of the AX.25 headers leaving the CTRL/PID bytes
217 */
218 skb_pull(skb, ax25_addr_size(&dp));
219
220 /* For our port addresses ? */
221 if (ax25cmp(&dest, dev_addr) == 0 && dp.lastrepeat + 1 == dp.ndigi)
222 mine = 1;
223
224 /* Also match on any registered callsign from L3/4 */
225 if (!mine && ax25_listen_mine(&dest, dev) && dp.lastrepeat + 1 == dp.ndigi)
226 mine = 1;
227
228 /* UI frame - bypass LAPB processing */
229 if ((*skb->data & ~0x10) == AX25_UI && dp.lastrepeat + 1 == dp.ndigi) {
230 skb_set_transport_header(skb, 2); /* skip control and pid */
231
232 ax25_send_to_raw(&dest, skb, skb->data[1]);
233
234 if (!mine && ax25cmp(&dest, (ax25_address *)dev->broadcast) != 0)
235 goto free;
236
237 /* Now we are pointing at the pid byte */
238 switch (skb->data[1]) {
239 case AX25_P_IP:
240 skb_pull(skb,2); /* drop PID/CTRL */
241 skb_reset_transport_header(skb);
242 skb_reset_network_header(skb);
243 skb->dev = dev;
244 skb->pkt_type = PACKET_HOST;
245 skb->protocol = htons(ETH_P_IP);
246 netif_rx(skb);
247 break;
248
249 case AX25_P_ARP:
250 skb_pull(skb,2);
251 skb_reset_transport_header(skb);
252 skb_reset_network_header(skb);
253 skb->dev = dev;
254 skb->pkt_type = PACKET_HOST;
255 skb->protocol = htons(ETH_P_ARP);
256 netif_rx(skb);
257 break;
258 case AX25_P_TEXT:
259 /* Now find a suitable dgram socket */
260 sk = ax25_get_socket(&dest, &src, SOCK_DGRAM);
261 if (sk != NULL) {
262 bh_lock_sock(sk);
263 if (atomic_read(&sk->sk_rmem_alloc) >=
264 sk->sk_rcvbuf) {
265 kfree_skb(skb);
266 } else {
267 /*
268 * Remove the control and PID.
269 */
270 skb_pull(skb, 2);
271 if (sock_queue_rcv_skb(sk, skb) != 0)
272 kfree_skb(skb);
273 }
274 bh_unlock_sock(sk);
275 sock_put(sk);
276 } else {
277 kfree_skb(skb);
278 }
279 break;
280
281 default:
282 kfree_skb(skb); /* Will scan SOCK_AX25 RAW sockets */
283 break;
284 }
285
286 return 0;
287 }
288
289 /*
290 * Is connected mode supported on this device ?
291 * If not, should we DM the incoming frame (except DMs) or
292 * silently ignore them. For now we stay quiet.
293 */
294 if (ax25_dev->values[AX25_VALUES_CONMODE] == 0)
295 goto free;
296
297 /* LAPB */
298
299 /* AX.25 state 1-4 */
300
301 ax25_digi_invert(&dp, &reverse_dp);
302
303 if ((ax25 = ax25_find_cb(&dest, &src, &reverse_dp, dev)) != NULL) {
304 /*
305 * Process the frame. If it is queued up internally it
306 * returns one otherwise we free it immediately. This
307 * routine itself wakes the user context layers so we do
308 * no further work
309 */
310 if (ax25_process_rx_frame(ax25, skb, type, dama) == 0)
311 kfree_skb(skb);
312
313 ax25_cb_put(ax25);
314 return 0;
315 }
316
317 /* AX.25 state 0 (disconnected) */
318
319 /* a) received not a SABM(E) */
320
321 if ((*skb->data & ~AX25_PF) != AX25_SABM &&
322 (*skb->data & ~AX25_PF) != AX25_SABME) {
323 /*
324 * Never reply to a DM. Also ignore any connects for
325 * addresses that are not our interfaces and not a socket.
326 */
327 if ((*skb->data & ~AX25_PF) != AX25_DM && mine)
328 ax25_return_dm(dev, &src, &dest, &dp);
329
330 goto free;
331 }
332
333 /* b) received SABM(E) */
334
335 if (dp.lastrepeat + 1 == dp.ndigi)
336 sk = ax25_find_listener(&dest, 0, dev, SOCK_SEQPACKET);
337 else
338 sk = ax25_find_listener(next_digi, 1, dev, SOCK_SEQPACKET);
339
340 if (sk != NULL) {
341 bh_lock_sock(sk);
342 if (sk_acceptq_is_full(sk) ||
343 (make = ax25_make_new(sk, ax25_dev)) == NULL) {
344 if (mine)
345 ax25_return_dm(dev, &src, &dest, &dp);
346 kfree_skb(skb);
347 bh_unlock_sock(sk);
348 sock_put(sk);
349
350 return 0;
351 }
352
353 ax25 = sk_to_ax25(make);
354 skb_set_owner_r(skb, make);
355 skb_queue_head(&sk->sk_receive_queue, skb);
356
357 make->sk_state = TCP_ESTABLISHED;
358
359 sk_acceptq_added(sk);
360 bh_unlock_sock(sk);
361 } else {
362 if (!mine)
363 goto free;
364
365 if ((ax25 = ax25_create_cb()) == NULL) {
366 ax25_return_dm(dev, &src, &dest, &dp);
367 goto free;
368 }
369
370 ax25_fillin_cb(ax25, ax25_dev);
371 }
372
373 ax25->source_addr = dest;
374 ax25->dest_addr = src;
375
376 /*
377 * Sort out any digipeated paths.
378 */
379 if (dp.ndigi && !ax25->digipeat &&
380 (ax25->digipeat = kmalloc(sizeof(ax25_digi), GFP_ATOMIC)) == NULL) {
381 kfree_skb(skb);
382 ax25_destroy_socket(ax25);
383 if (sk)
384 sock_put(sk);
385 return 0;
386 }
387
388 if (dp.ndigi == 0) {
389 kfree(ax25->digipeat);
390 ax25->digipeat = NULL;
391 } else {
392 /* Reverse the source SABM's path */
393 memcpy(ax25->digipeat, &reverse_dp, sizeof(ax25_digi));
394 }
395
396 if ((*skb->data & ~AX25_PF) == AX25_SABME) {
397 ax25->modulus = AX25_EMODULUS;
398 ax25->window = ax25_dev->values[AX25_VALUES_EWINDOW];
399 } else {
400 ax25->modulus = AX25_MODULUS;
401 ax25->window = ax25_dev->values[AX25_VALUES_WINDOW];
402 }
403
404 ax25_send_control(ax25, AX25_UA, AX25_POLLON, AX25_RESPONSE);
405
406 #ifdef CONFIG_AX25_DAMA_SLAVE
407 if (dama && ax25->ax25_dev->values[AX25_VALUES_PROTOCOL] == AX25_PROTO_DAMA_SLAVE)
408 ax25_dama_on(ax25);
409 #endif
410
411 ax25->state = AX25_STATE_3;
412
413 ax25_cb_add(ax25);
414
415 ax25_start_heartbeat(ax25);
416 ax25_start_t3timer(ax25);
417 ax25_start_idletimer(ax25);
418
419 if (sk) {
420 if (!sock_flag(sk, SOCK_DEAD))
421 sk->sk_data_ready(sk);
422 sock_put(sk);
423 } else {
424 free:
425 kfree_skb(skb);
426 }
427 return 0;
428 }
429
430 /*
431 * Receive an AX.25 frame via a SLIP interface.
432 */
ax25_kiss_rcv(struct sk_buff * skb,struct net_device * dev,struct packet_type * ptype,struct net_device * orig_dev)433 int ax25_kiss_rcv(struct sk_buff *skb, struct net_device *dev,
434 struct packet_type *ptype, struct net_device *orig_dev)
435 {
436 skb_orphan(skb);
437
438 if (!net_eq(dev_net(dev), &init_net)) {
439 kfree_skb(skb);
440 return 0;
441 }
442
443 if ((*skb->data & 0x0F) != 0) {
444 kfree_skb(skb); /* Not a KISS data frame */
445 return 0;
446 }
447
448 skb_pull(skb, AX25_KISS_HEADER_LEN); /* Remove the KISS byte */
449
450 return ax25_rcv(skb, dev, (const ax25_address *)dev->dev_addr, ptype);
451 }
452