Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: GPL-2.0
2// Copyright (c) 2010-2011 EIA Electronics,
3// Pieter Beyens <pieter.beyens@eia.be>
4// Copyright (c) 2010-2011 EIA Electronics,
5// Kurt Van Dijck <kurt.van.dijck@eia.be>
6// Copyright (c) 2018 Protonic,
7// Robin van der Gracht <robin@protonic.nl>
8// Copyright (c) 2017-2019 Pengutronix,
9// Marc Kleine-Budde <kernel@pengutronix.de>
10// Copyright (c) 2017-2019 Pengutronix,
11// Oleksij Rempel <kernel@pengutronix.de>
12
13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15#include <linux/can/can-ml.h>
16#include <linux/can/core.h>
17#include <linux/can/skb.h>
18#include <linux/errqueue.h>
19#include <linux/if_arp.h>
20
21#include "j1939-priv.h"
22
23#define J1939_MIN_NAMELEN CAN_REQUIRED_SIZE(struct sockaddr_can, can_addr.j1939)
24
25/* conversion function between struct sock::sk_priority from linux and
26 * j1939 priority field
27 */
28static inline priority_t j1939_prio(u32 sk_priority)
29{
30 sk_priority = min(sk_priority, 7U);
31
32 return 7 - sk_priority;
33}
34
35static inline u32 j1939_to_sk_priority(priority_t prio)
36{
37 return 7 - prio;
38}
39
40/* function to see if pgn is to be evaluated */
41static inline bool j1939_pgn_is_valid(pgn_t pgn)
42{
43 return pgn <= J1939_PGN_MAX;
44}
45
46/* test function to avoid non-zero DA placeholder for pdu1 pgn's */
47static inline bool j1939_pgn_is_clean_pdu(pgn_t pgn)
48{
49 if (j1939_pgn_is_pdu1(pgn))
50 return !(pgn & 0xff);
51 else
52 return true;
53}
54
55static inline void j1939_sock_pending_add(struct sock *sk)
56{
57 struct j1939_sock *jsk = j1939_sk(sk);
58
59 atomic_inc(&jsk->skb_pending);
60}
61
62static int j1939_sock_pending_get(struct sock *sk)
63{
64 struct j1939_sock *jsk = j1939_sk(sk);
65
66 return atomic_read(&jsk->skb_pending);
67}
68
69void j1939_sock_pending_del(struct sock *sk)
70{
71 struct j1939_sock *jsk = j1939_sk(sk);
72
73 /* atomic_dec_return returns the new value */
74 if (!atomic_dec_return(&jsk->skb_pending))
75 wake_up(&jsk->waitq); /* no pending SKB's */
76}
77
78static void j1939_jsk_add(struct j1939_priv *priv, struct j1939_sock *jsk)
79{
80 jsk->state |= J1939_SOCK_BOUND;
81 j1939_priv_get(priv);
82
83 write_lock_bh(&priv->j1939_socks_lock);
84 list_add_tail(&jsk->list, &priv->j1939_socks);
85 write_unlock_bh(&priv->j1939_socks_lock);
86}
87
88static void j1939_jsk_del(struct j1939_priv *priv, struct j1939_sock *jsk)
89{
90 write_lock_bh(&priv->j1939_socks_lock);
91 list_del_init(&jsk->list);
92 write_unlock_bh(&priv->j1939_socks_lock);
93
94 j1939_priv_put(priv);
95 jsk->state &= ~J1939_SOCK_BOUND;
96}
97
98static bool j1939_sk_queue_session(struct j1939_session *session)
99{
100 struct j1939_sock *jsk = j1939_sk(session->sk);
101 bool empty;
102
103 spin_lock_bh(&jsk->sk_session_queue_lock);
104 empty = list_empty(&jsk->sk_session_queue);
105 j1939_session_get(session);
106 list_add_tail(&session->sk_session_queue_entry, &jsk->sk_session_queue);
107 spin_unlock_bh(&jsk->sk_session_queue_lock);
108 j1939_sock_pending_add(&jsk->sk);
109
110 return empty;
111}
112
113static struct
114j1939_session *j1939_sk_get_incomplete_session(struct j1939_sock *jsk)
115{
116 struct j1939_session *session = NULL;
117
118 spin_lock_bh(&jsk->sk_session_queue_lock);
119 if (!list_empty(&jsk->sk_session_queue)) {
120 session = list_last_entry(&jsk->sk_session_queue,
121 struct j1939_session,
122 sk_session_queue_entry);
123 if (session->total_queued_size == session->total_message_size)
124 session = NULL;
125 else
126 j1939_session_get(session);
127 }
128 spin_unlock_bh(&jsk->sk_session_queue_lock);
129
130 return session;
131}
132
133static void j1939_sk_queue_drop_all(struct j1939_priv *priv,
134 struct j1939_sock *jsk, int err)
135{
136 struct j1939_session *session, *tmp;
137
138 netdev_dbg(priv->ndev, "%s: err: %i\n", __func__, err);
139 spin_lock_bh(&jsk->sk_session_queue_lock);
140 list_for_each_entry_safe(session, tmp, &jsk->sk_session_queue,
141 sk_session_queue_entry) {
142 list_del_init(&session->sk_session_queue_entry);
143 session->err = err;
144 j1939_session_put(session);
145 }
146 spin_unlock_bh(&jsk->sk_session_queue_lock);
147}
148
149static void j1939_sk_queue_activate_next_locked(struct j1939_session *session)
150{
151 struct j1939_sock *jsk;
152 struct j1939_session *first;
153 int err;
154
155 /* RX-Session don't have a socket (yet) */
156 if (!session->sk)
157 return;
158
159 jsk = j1939_sk(session->sk);
160 lockdep_assert_held(&jsk->sk_session_queue_lock);
161
162 err = session->err;
163
164 first = list_first_entry_or_null(&jsk->sk_session_queue,
165 struct j1939_session,
166 sk_session_queue_entry);
167
168 /* Some else has already activated the next session */
169 if (first != session)
170 return;
171
172activate_next:
173 list_del_init(&first->sk_session_queue_entry);
174 j1939_session_put(first);
175 first = list_first_entry_or_null(&jsk->sk_session_queue,
176 struct j1939_session,
177 sk_session_queue_entry);
178 if (!first)
179 return;
180
181 if (j1939_session_activate(first)) {
182 netdev_warn_once(first->priv->ndev,
183 "%s: 0x%p: Identical session is already activated.\n",
184 __func__, first);
185 first->err = -EBUSY;
186 goto activate_next;
187 } else {
188 /* Give receiver some time (arbitrary chosen) to recover */
189 int time_ms = 0;
190
191 if (err)
192 time_ms = 10 + get_random_u32_below(16);
193
194 j1939_tp_schedule_txtimer(first, time_ms);
195 }
196}
197
198void j1939_sk_queue_activate_next(struct j1939_session *session)
199{
200 struct j1939_sock *jsk;
201
202 if (!session->sk)
203 return;
204
205 jsk = j1939_sk(session->sk);
206
207 spin_lock_bh(&jsk->sk_session_queue_lock);
208 j1939_sk_queue_activate_next_locked(session);
209 spin_unlock_bh(&jsk->sk_session_queue_lock);
210}
211
212static bool j1939_sk_match_dst(struct j1939_sock *jsk,
213 const struct j1939_sk_buff_cb *skcb)
214{
215 if ((jsk->state & J1939_SOCK_PROMISC))
216 return true;
217
218 /* Destination address filter */
219 if (jsk->addr.src_name && skcb->addr.dst_name) {
220 if (jsk->addr.src_name != skcb->addr.dst_name)
221 return false;
222 } else {
223 /* receive (all sockets) if
224 * - all packages that match our bind() address
225 * - all broadcast on a socket if SO_BROADCAST
226 * is set
227 */
228 if (j1939_address_is_unicast(skcb->addr.da)) {
229 if (jsk->addr.sa != skcb->addr.da)
230 return false;
231 } else if (!sock_flag(&jsk->sk, SOCK_BROADCAST)) {
232 /* receiving broadcast without SO_BROADCAST
233 * flag is not allowed
234 */
235 return false;
236 }
237 }
238
239 /* Source address filter */
240 if (jsk->state & J1939_SOCK_CONNECTED) {
241 /* receive (all sockets) if
242 * - all packages that match our connect() name or address
243 */
244 if (jsk->addr.dst_name && skcb->addr.src_name) {
245 if (jsk->addr.dst_name != skcb->addr.src_name)
246 return false;
247 } else {
248 if (jsk->addr.da != skcb->addr.sa)
249 return false;
250 }
251 }
252
253 /* PGN filter */
254 if (j1939_pgn_is_valid(jsk->pgn_rx_filter) &&
255 jsk->pgn_rx_filter != skcb->addr.pgn)
256 return false;
257
258 return true;
259}
260
261/* matches skb control buffer (addr) with a j1939 filter */
262static bool j1939_sk_match_filter(struct j1939_sock *jsk,
263 const struct j1939_sk_buff_cb *skcb)
264{
265 const struct j1939_filter *f;
266 int nfilter;
267
268 spin_lock_bh(&jsk->filters_lock);
269
270 f = jsk->filters;
271 nfilter = jsk->nfilters;
272
273 if (!nfilter)
274 /* receive all when no filters are assigned */
275 goto filter_match_found;
276
277 for (; nfilter; ++f, --nfilter) {
278 if ((skcb->addr.pgn & f->pgn_mask) != f->pgn)
279 continue;
280 if ((skcb->addr.sa & f->addr_mask) != f->addr)
281 continue;
282 if ((skcb->addr.src_name & f->name_mask) != f->name)
283 continue;
284 goto filter_match_found;
285 }
286
287 spin_unlock_bh(&jsk->filters_lock);
288 return false;
289
290filter_match_found:
291 spin_unlock_bh(&jsk->filters_lock);
292 return true;
293}
294
295static bool j1939_sk_recv_match_one(struct j1939_sock *jsk,
296 const struct j1939_sk_buff_cb *skcb)
297{
298 if (!(jsk->state & J1939_SOCK_BOUND))
299 return false;
300
301 if (!j1939_sk_match_dst(jsk, skcb))
302 return false;
303
304 if (!j1939_sk_match_filter(jsk, skcb))
305 return false;
306
307 return true;
308}
309
310static void j1939_sk_recv_one(struct j1939_sock *jsk, struct sk_buff *oskb)
311{
312 const struct j1939_sk_buff_cb *oskcb = j1939_skb_to_cb(oskb);
313 struct j1939_sk_buff_cb *skcb;
314 enum skb_drop_reason reason;
315 struct sk_buff *skb;
316
317 if (oskb->sk == &jsk->sk)
318 return;
319
320 if (!j1939_sk_recv_match_one(jsk, oskcb))
321 return;
322
323 skb = skb_clone(oskb, GFP_ATOMIC);
324 if (!skb) {
325 pr_warn("skb clone failed\n");
326 return;
327 }
328 can_skb_set_owner(skb, oskb->sk);
329
330 skcb = j1939_skb_to_cb(skb);
331 skcb->msg_flags &= ~(MSG_DONTROUTE);
332 if (skb->sk)
333 skcb->msg_flags |= MSG_DONTROUTE;
334
335 if (sock_queue_rcv_skb_reason(&jsk->sk, skb, &reason) < 0)
336 sk_skb_reason_drop(&jsk->sk, skb, reason);
337}
338
339bool j1939_sk_recv_match(struct j1939_priv *priv, struct j1939_sk_buff_cb *skcb)
340{
341 struct j1939_sock *jsk;
342 bool match = false;
343
344 read_lock_bh(&priv->j1939_socks_lock);
345 list_for_each_entry(jsk, &priv->j1939_socks, list) {
346 match = j1939_sk_recv_match_one(jsk, skcb);
347 if (match)
348 break;
349 }
350 read_unlock_bh(&priv->j1939_socks_lock);
351
352 return match;
353}
354
355void j1939_sk_recv(struct j1939_priv *priv, struct sk_buff *skb)
356{
357 struct j1939_sock *jsk;
358
359 read_lock_bh(&priv->j1939_socks_lock);
360 list_for_each_entry(jsk, &priv->j1939_socks, list) {
361 j1939_sk_recv_one(jsk, skb);
362 }
363 read_unlock_bh(&priv->j1939_socks_lock);
364}
365
366static void j1939_sk_sock_destruct(struct sock *sk)
367{
368 struct j1939_sock *jsk = j1939_sk(sk);
369
370 /* This function will be called by the generic networking code, when
371 * the socket is ultimately closed (sk->sk_destruct).
372 *
373 * The race between
374 * - processing a received CAN frame
375 * (can_receive -> j1939_can_recv)
376 * and accessing j1939_priv
377 * ... and ...
378 * - closing a socket
379 * (j1939_can_rx_unregister -> can_rx_unregister)
380 * and calling the final j1939_priv_put()
381 *
382 * is avoided by calling the final j1939_priv_put() from this
383 * RCU deferred cleanup call.
384 */
385 if (jsk->priv) {
386 j1939_priv_put(jsk->priv);
387 jsk->priv = NULL;
388 }
389
390 /* call generic CAN sock destruct */
391 can_sock_destruct(sk);
392}
393
394static int j1939_sk_init(struct sock *sk)
395{
396 struct j1939_sock *jsk = j1939_sk(sk);
397
398 /* Ensure that "sk" is first member in "struct j1939_sock", so that we
399 * can skip it during memset().
400 */
401 BUILD_BUG_ON(offsetof(struct j1939_sock, sk) != 0);
402 memset((void *)jsk + sizeof(jsk->sk), 0x0,
403 sizeof(*jsk) - sizeof(jsk->sk));
404
405 INIT_LIST_HEAD(&jsk->list);
406 init_waitqueue_head(&jsk->waitq);
407 jsk->sk.sk_priority = j1939_to_sk_priority(6);
408 jsk->sk.sk_reuse = 1; /* per default */
409 jsk->addr.sa = J1939_NO_ADDR;
410 jsk->addr.da = J1939_NO_ADDR;
411 jsk->addr.pgn = J1939_NO_PGN;
412 jsk->pgn_rx_filter = J1939_NO_PGN;
413 atomic_set(&jsk->skb_pending, 0);
414 spin_lock_init(&jsk->sk_session_queue_lock);
415 INIT_LIST_HEAD(&jsk->sk_session_queue);
416 spin_lock_init(&jsk->filters_lock);
417
418 /* j1939_sk_sock_destruct() depends on SOCK_RCU_FREE flag */
419 sock_set_flag(sk, SOCK_RCU_FREE);
420 sk->sk_destruct = j1939_sk_sock_destruct;
421 sk->sk_protocol = CAN_J1939;
422
423 return 0;
424}
425
426static int j1939_sk_sanity_check(struct sockaddr_can *addr, int len)
427{
428 if (!addr)
429 return -EDESTADDRREQ;
430 if (len < J1939_MIN_NAMELEN)
431 return -EINVAL;
432 if (addr->can_family != AF_CAN)
433 return -EINVAL;
434 if (!addr->can_ifindex)
435 return -ENODEV;
436 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) &&
437 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn))
438 return -EINVAL;
439
440 return 0;
441}
442
443static int j1939_sk_bind(struct socket *sock, struct sockaddr_unsized *uaddr, int len)
444{
445 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
446 struct j1939_sock *jsk = j1939_sk(sock->sk);
447 struct j1939_priv *priv;
448 struct sock *sk;
449 struct net *net;
450 int ret = 0;
451
452 ret = j1939_sk_sanity_check(addr, len);
453 if (ret)
454 return ret;
455
456 lock_sock(sock->sk);
457
458 priv = jsk->priv;
459 sk = sock->sk;
460 net = sock_net(sk);
461
462 /* Already bound to an interface? */
463 if (jsk->state & J1939_SOCK_BOUND) {
464 /* A re-bind() to a different interface is not
465 * supported.
466 */
467 if (jsk->ifindex != addr->can_ifindex) {
468 ret = -EINVAL;
469 goto out_release_sock;
470 }
471
472 /* drop old references */
473 j1939_jsk_del(priv, jsk);
474 j1939_local_ecu_put(priv, jsk->addr.src_name, jsk->addr.sa);
475 } else {
476 struct can_ml_priv *can_ml;
477 struct net_device *ndev;
478
479 ndev = dev_get_by_index(net, addr->can_ifindex);
480 if (!ndev) {
481 ret = -ENODEV;
482 goto out_release_sock;
483 }
484
485 if (ndev->reg_state != NETREG_REGISTERED) {
486 dev_put(ndev);
487 ret = -ENODEV;
488 goto out_release_sock;
489 }
490
491 can_ml = can_get_ml_priv(ndev);
492 if (!can_ml) {
493 dev_put(ndev);
494 ret = -ENODEV;
495 goto out_release_sock;
496 }
497
498 if (!(ndev->flags & IFF_UP)) {
499 dev_put(ndev);
500 ret = -ENETDOWN;
501 goto out_release_sock;
502 }
503
504 priv = j1939_netdev_start(ndev);
505 dev_put(ndev);
506 if (IS_ERR(priv)) {
507 ret = PTR_ERR(priv);
508 goto out_release_sock;
509 }
510
511 jsk->ifindex = addr->can_ifindex;
512
513 /* the corresponding j1939_priv_put() is called via
514 * sk->sk_destruct, which points to j1939_sk_sock_destruct()
515 */
516 j1939_priv_get(priv);
517 jsk->priv = priv;
518 }
519
520 /* set default transmit pgn */
521 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
522 jsk->pgn_rx_filter = addr->can_addr.j1939.pgn;
523 jsk->addr.src_name = addr->can_addr.j1939.name;
524 jsk->addr.sa = addr->can_addr.j1939.addr;
525
526 /* get new references */
527 ret = j1939_local_ecu_get(priv, jsk->addr.src_name, jsk->addr.sa);
528 if (ret) {
529 j1939_netdev_stop(priv);
530 jsk->priv = NULL;
531 synchronize_rcu();
532 j1939_priv_put(priv);
533 goto out_release_sock;
534 }
535
536 j1939_jsk_add(priv, jsk);
537
538 out_release_sock: /* fall through */
539 release_sock(sock->sk);
540
541 return ret;
542}
543
544static int j1939_sk_connect(struct socket *sock, struct sockaddr_unsized *uaddr,
545 int len, int flags)
546{
547 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
548 struct j1939_sock *jsk = j1939_sk(sock->sk);
549 int ret = 0;
550
551 ret = j1939_sk_sanity_check(addr, len);
552 if (ret)
553 return ret;
554
555 lock_sock(sock->sk);
556
557 /* bind() before connect() is mandatory */
558 if (!(jsk->state & J1939_SOCK_BOUND)) {
559 ret = -EINVAL;
560 goto out_release_sock;
561 }
562
563 /* A connect() to a different interface is not supported. */
564 if (jsk->ifindex != addr->can_ifindex) {
565 ret = -EINVAL;
566 goto out_release_sock;
567 }
568
569 if (!addr->can_addr.j1939.name &&
570 addr->can_addr.j1939.addr == J1939_NO_ADDR &&
571 !sock_flag(&jsk->sk, SOCK_BROADCAST)) {
572 /* broadcast, but SO_BROADCAST not set */
573 ret = -EACCES;
574 goto out_release_sock;
575 }
576
577 jsk->addr.dst_name = addr->can_addr.j1939.name;
578 jsk->addr.da = addr->can_addr.j1939.addr;
579
580 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
581 jsk->addr.pgn = addr->can_addr.j1939.pgn;
582
583 jsk->state |= J1939_SOCK_CONNECTED;
584
585 out_release_sock: /* fall through */
586 release_sock(sock->sk);
587
588 return ret;
589}
590
591static void j1939_sk_sock2sockaddr_can(struct sockaddr_can *addr,
592 const struct j1939_sock *jsk, int peer)
593{
594 /* There are two holes (2 bytes and 3 bytes) to clear to avoid
595 * leaking kernel information to user space.
596 */
597 memset(addr, 0, J1939_MIN_NAMELEN);
598
599 addr->can_family = AF_CAN;
600 addr->can_ifindex = jsk->ifindex;
601 addr->can_addr.j1939.pgn = jsk->addr.pgn;
602 if (peer) {
603 addr->can_addr.j1939.name = jsk->addr.dst_name;
604 addr->can_addr.j1939.addr = jsk->addr.da;
605 } else {
606 addr->can_addr.j1939.name = jsk->addr.src_name;
607 addr->can_addr.j1939.addr = jsk->addr.sa;
608 }
609}
610
611static int j1939_sk_getname(struct socket *sock, struct sockaddr *uaddr,
612 int peer)
613{
614 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
615 struct sock *sk = sock->sk;
616 struct j1939_sock *jsk = j1939_sk(sk);
617 int ret = 0;
618
619 lock_sock(sk);
620
621 if (peer && !(jsk->state & J1939_SOCK_CONNECTED)) {
622 ret = -EADDRNOTAVAIL;
623 goto failure;
624 }
625
626 j1939_sk_sock2sockaddr_can(addr, jsk, peer);
627 ret = J1939_MIN_NAMELEN;
628
629 failure:
630 release_sock(sk);
631
632 return ret;
633}
634
635static int j1939_sk_release(struct socket *sock)
636{
637 struct sock *sk = sock->sk;
638 struct j1939_sock *jsk;
639
640 if (!sk)
641 return 0;
642
643 lock_sock(sk);
644 jsk = j1939_sk(sk);
645
646 if (jsk->state & J1939_SOCK_BOUND) {
647 struct j1939_priv *priv = jsk->priv;
648
649 if (wait_event_interruptible(jsk->waitq,
650 !j1939_sock_pending_get(&jsk->sk))) {
651 j1939_cancel_active_session(priv, sk);
652 j1939_sk_queue_drop_all(priv, jsk, ESHUTDOWN);
653 }
654
655 j1939_jsk_del(priv, jsk);
656
657 j1939_local_ecu_put(priv, jsk->addr.src_name,
658 jsk->addr.sa);
659
660 j1939_netdev_stop(priv);
661 }
662
663 kfree(jsk->filters);
664 sock_orphan(sk);
665 sock->sk = NULL;
666
667 release_sock(sk);
668 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1);
669 sock_put(sk);
670
671 return 0;
672}
673
674static int j1939_sk_setsockopt_flag(struct j1939_sock *jsk, sockptr_t optval,
675 unsigned int optlen, int flag)
676{
677 int tmp;
678
679 if (optlen != sizeof(tmp))
680 return -EINVAL;
681 if (copy_from_sockptr(&tmp, optval, optlen))
682 return -EFAULT;
683 lock_sock(&jsk->sk);
684 if (tmp)
685 jsk->state |= flag;
686 else
687 jsk->state &= ~flag;
688 release_sock(&jsk->sk);
689 return tmp;
690}
691
692static int j1939_sk_setsockopt(struct socket *sock, int level, int optname,
693 sockptr_t optval, unsigned int optlen)
694{
695 struct sock *sk = sock->sk;
696 struct j1939_sock *jsk = j1939_sk(sk);
697 int tmp, count = 0, ret = 0;
698 struct j1939_filter *filters = NULL, *ofilters;
699
700 if (level != SOL_CAN_J1939)
701 return -EINVAL;
702
703 switch (optname) {
704 case SO_J1939_FILTER:
705 if (!sockptr_is_null(optval) && optlen != 0) {
706 struct j1939_filter *f;
707 int c;
708
709 if (optlen % sizeof(*filters) != 0)
710 return -EINVAL;
711
712 if (optlen > J1939_FILTER_MAX *
713 sizeof(struct j1939_filter))
714 return -EINVAL;
715
716 count = optlen / sizeof(*filters);
717 filters = memdup_sockptr(optval, optlen);
718 if (IS_ERR(filters))
719 return PTR_ERR(filters);
720
721 for (f = filters, c = count; c; f++, c--) {
722 f->name &= f->name_mask;
723 f->pgn &= f->pgn_mask;
724 f->addr &= f->addr_mask;
725 }
726 }
727
728 lock_sock(&jsk->sk);
729 spin_lock_bh(&jsk->filters_lock);
730 ofilters = jsk->filters;
731 jsk->filters = filters;
732 jsk->nfilters = count;
733 spin_unlock_bh(&jsk->filters_lock);
734 release_sock(&jsk->sk);
735 kfree(ofilters);
736 return 0;
737 case SO_J1939_PROMISC:
738 return j1939_sk_setsockopt_flag(jsk, optval, optlen,
739 J1939_SOCK_PROMISC);
740 case SO_J1939_ERRQUEUE:
741 ret = j1939_sk_setsockopt_flag(jsk, optval, optlen,
742 J1939_SOCK_ERRQUEUE);
743 if (ret < 0)
744 return ret;
745
746 if (!(jsk->state & J1939_SOCK_ERRQUEUE))
747 skb_queue_purge(&sk->sk_error_queue);
748 return ret;
749 case SO_J1939_SEND_PRIO:
750 if (optlen != sizeof(tmp))
751 return -EINVAL;
752 if (copy_from_sockptr(&tmp, optval, optlen))
753 return -EFAULT;
754 if (tmp < 0 || tmp > 7)
755 return -EDOM;
756 if (tmp < 2 && !capable(CAP_NET_ADMIN))
757 return -EPERM;
758 lock_sock(&jsk->sk);
759 jsk->sk.sk_priority = j1939_to_sk_priority(tmp);
760 release_sock(&jsk->sk);
761 return 0;
762 default:
763 return -ENOPROTOOPT;
764 }
765}
766
767static int j1939_sk_getsockopt(struct socket *sock, int level, int optname,
768 char __user *optval, int __user *optlen)
769{
770 struct sock *sk = sock->sk;
771 struct j1939_sock *jsk = j1939_sk(sk);
772 int ret, ulen;
773 /* set defaults for using 'int' properties */
774 int tmp = 0;
775 int len = sizeof(tmp);
776 void *val = &tmp;
777
778 if (level != SOL_CAN_J1939)
779 return -EINVAL;
780 if (get_user(ulen, optlen))
781 return -EFAULT;
782 if (ulen < 0)
783 return -EINVAL;
784
785 lock_sock(&jsk->sk);
786 switch (optname) {
787 case SO_J1939_PROMISC:
788 tmp = (jsk->state & J1939_SOCK_PROMISC) ? 1 : 0;
789 break;
790 case SO_J1939_ERRQUEUE:
791 tmp = (jsk->state & J1939_SOCK_ERRQUEUE) ? 1 : 0;
792 break;
793 case SO_J1939_SEND_PRIO:
794 tmp = j1939_prio(jsk->sk.sk_priority);
795 break;
796 default:
797 ret = -ENOPROTOOPT;
798 goto no_copy;
799 }
800
801 /* copy to user, based on 'len' & 'val'
802 * but most sockopt's are 'int' properties, and have 'len' & 'val'
803 * left unchanged, but instead modified 'tmp'
804 */
805 if (len > ulen)
806 ret = -EFAULT;
807 else if (put_user(len, optlen))
808 ret = -EFAULT;
809 else if (copy_to_user(optval, val, len))
810 ret = -EFAULT;
811 else
812 ret = 0;
813 no_copy:
814 release_sock(&jsk->sk);
815 return ret;
816}
817
818static int j1939_sk_recvmsg(struct socket *sock, struct msghdr *msg,
819 size_t size, int flags)
820{
821 struct sock *sk = sock->sk;
822 struct sk_buff *skb;
823 struct j1939_sk_buff_cb *skcb;
824 int ret = 0;
825
826 if (flags & ~(MSG_DONTWAIT | MSG_ERRQUEUE | MSG_CMSG_COMPAT))
827 return -EINVAL;
828
829 if (flags & MSG_ERRQUEUE)
830 return sock_recv_errqueue(sock->sk, msg, size, SOL_CAN_J1939,
831 SCM_J1939_ERRQUEUE);
832
833 skb = skb_recv_datagram(sk, flags, &ret);
834 if (!skb)
835 return ret;
836
837 if (size < skb->len)
838 msg->msg_flags |= MSG_TRUNC;
839 else
840 size = skb->len;
841
842 ret = memcpy_to_msg(msg, skb->data, size);
843 if (ret < 0) {
844 skb_free_datagram(sk, skb);
845 return ret;
846 }
847
848 skcb = j1939_skb_to_cb(skb);
849 if (j1939_address_is_valid(skcb->addr.da))
850 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_ADDR,
851 sizeof(skcb->addr.da), &skcb->addr.da);
852
853 if (skcb->addr.dst_name)
854 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_NAME,
855 sizeof(skcb->addr.dst_name), &skcb->addr.dst_name);
856
857 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_PRIO,
858 sizeof(skcb->priority), &skcb->priority);
859
860 if (msg->msg_name) {
861 struct sockaddr_can *paddr = msg->msg_name;
862
863 msg->msg_namelen = J1939_MIN_NAMELEN;
864 memset(msg->msg_name, 0, msg->msg_namelen);
865 paddr->can_family = AF_CAN;
866 paddr->can_ifindex = skb->skb_iif;
867 paddr->can_addr.j1939.name = skcb->addr.src_name;
868 paddr->can_addr.j1939.addr = skcb->addr.sa;
869 paddr->can_addr.j1939.pgn = skcb->addr.pgn;
870 }
871
872 sock_recv_cmsgs(msg, sk, skb);
873 msg->msg_flags |= skcb->msg_flags;
874 skb_free_datagram(sk, skb);
875
876 return size;
877}
878
879static struct sk_buff *j1939_sk_alloc_skb(struct net_device *ndev,
880 struct sock *sk,
881 struct msghdr *msg, size_t size,
882 int *errcode)
883{
884 struct j1939_sock *jsk = j1939_sk(sk);
885 struct j1939_sk_buff_cb *skcb;
886 struct sk_buff *skb;
887 int ret;
888
889 skb = sock_alloc_send_skb(sk,
890 size +
891 sizeof(struct can_frame) -
892 sizeof(((struct can_frame *)NULL)->data) +
893 sizeof(struct can_skb_priv),
894 msg->msg_flags & MSG_DONTWAIT, &ret);
895 if (!skb)
896 goto failure;
897
898 can_skb_reserve(skb);
899 can_skb_prv(skb)->ifindex = ndev->ifindex;
900 can_skb_prv(skb)->skbcnt = 0;
901 skb_reserve(skb, offsetof(struct can_frame, data));
902
903 ret = memcpy_from_msg(skb_put(skb, size), msg, size);
904 if (ret < 0)
905 goto free_skb;
906
907 skb->dev = ndev;
908
909 skcb = j1939_skb_to_cb(skb);
910 memset(skcb, 0, sizeof(*skcb));
911 skcb->addr = jsk->addr;
912 skcb->priority = j1939_prio(READ_ONCE(sk->sk_priority));
913
914 if (msg->msg_name) {
915 struct sockaddr_can *addr = msg->msg_name;
916
917 if (addr->can_addr.j1939.name ||
918 addr->can_addr.j1939.addr != J1939_NO_ADDR) {
919 skcb->addr.dst_name = addr->can_addr.j1939.name;
920 skcb->addr.da = addr->can_addr.j1939.addr;
921 }
922 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
923 skcb->addr.pgn = addr->can_addr.j1939.pgn;
924 }
925
926 *errcode = ret;
927 return skb;
928
929free_skb:
930 kfree_skb(skb);
931failure:
932 *errcode = ret;
933 return NULL;
934}
935
936static size_t j1939_sk_opt_stats_get_size(enum j1939_sk_errqueue_type type)
937{
938 switch (type) {
939 case J1939_ERRQUEUE_RX_RTS:
940 return
941 nla_total_size(sizeof(u32)) + /* J1939_NLA_TOTAL_SIZE */
942 nla_total_size(sizeof(u32)) + /* J1939_NLA_PGN */
943 nla_total_size(sizeof(u64)) + /* J1939_NLA_SRC_NAME */
944 nla_total_size(sizeof(u64)) + /* J1939_NLA_DEST_NAME */
945 nla_total_size(sizeof(u8)) + /* J1939_NLA_SRC_ADDR */
946 nla_total_size(sizeof(u8)) + /* J1939_NLA_DEST_ADDR */
947 0;
948 default:
949 return
950 nla_total_size(sizeof(u32)) + /* J1939_NLA_BYTES_ACKED */
951 0;
952 }
953}
954
955static struct sk_buff *
956j1939_sk_get_timestamping_opt_stats(struct j1939_session *session,
957 enum j1939_sk_errqueue_type type)
958{
959 struct sk_buff *stats;
960 u32 size;
961
962 stats = alloc_skb(j1939_sk_opt_stats_get_size(type), GFP_ATOMIC);
963 if (!stats)
964 return NULL;
965
966 if (session->skcb.addr.type == J1939_SIMPLE)
967 size = session->total_message_size;
968 else
969 size = min(session->pkt.tx_acked * 7,
970 session->total_message_size);
971
972 switch (type) {
973 case J1939_ERRQUEUE_RX_RTS:
974 nla_put_u32(stats, J1939_NLA_TOTAL_SIZE,
975 session->total_message_size);
976 nla_put_u32(stats, J1939_NLA_PGN,
977 session->skcb.addr.pgn);
978 nla_put_u64_64bit(stats, J1939_NLA_SRC_NAME,
979 session->skcb.addr.src_name, J1939_NLA_PAD);
980 nla_put_u64_64bit(stats, J1939_NLA_DEST_NAME,
981 session->skcb.addr.dst_name, J1939_NLA_PAD);
982 nla_put_u8(stats, J1939_NLA_SRC_ADDR,
983 session->skcb.addr.sa);
984 nla_put_u8(stats, J1939_NLA_DEST_ADDR,
985 session->skcb.addr.da);
986 break;
987 default:
988 nla_put_u32(stats, J1939_NLA_BYTES_ACKED, size);
989 }
990
991 return stats;
992}
993
994static void __j1939_sk_errqueue(struct j1939_session *session, struct sock *sk,
995 enum j1939_sk_errqueue_type type)
996{
997 struct j1939_priv *priv = session->priv;
998 struct j1939_sock *jsk;
999 struct sock_exterr_skb *serr;
1000 struct sk_buff *skb;
1001 char *state = "UNK";
1002 u32 tsflags;
1003 int err;
1004
1005 jsk = j1939_sk(sk);
1006
1007 if (!(jsk->state & J1939_SOCK_ERRQUEUE))
1008 return;
1009
1010 tsflags = READ_ONCE(sk->sk_tsflags);
1011 switch (type) {
1012 case J1939_ERRQUEUE_TX_ACK:
1013 if (!(tsflags & SOF_TIMESTAMPING_TX_ACK))
1014 return;
1015 break;
1016 case J1939_ERRQUEUE_TX_SCHED:
1017 if (!(tsflags & SOF_TIMESTAMPING_TX_SCHED))
1018 return;
1019 break;
1020 case J1939_ERRQUEUE_TX_ABORT:
1021 break;
1022 case J1939_ERRQUEUE_RX_RTS:
1023 fallthrough;
1024 case J1939_ERRQUEUE_RX_DPO:
1025 fallthrough;
1026 case J1939_ERRQUEUE_RX_ABORT:
1027 if (!(tsflags & SOF_TIMESTAMPING_RX_SOFTWARE))
1028 return;
1029 break;
1030 default:
1031 netdev_err(priv->ndev, "Unknown errqueue type %i\n", type);
1032 }
1033
1034 skb = j1939_sk_get_timestamping_opt_stats(session, type);
1035 if (!skb)
1036 return;
1037
1038 skb->tstamp = ktime_get_real();
1039
1040 BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb));
1041
1042 serr = SKB_EXT_ERR(skb);
1043 memset(serr, 0, sizeof(*serr));
1044 switch (type) {
1045 case J1939_ERRQUEUE_TX_ACK:
1046 serr->ee.ee_errno = ENOMSG;
1047 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING;
1048 serr->ee.ee_info = SCM_TSTAMP_ACK;
1049 state = "TX ACK";
1050 break;
1051 case J1939_ERRQUEUE_TX_SCHED:
1052 serr->ee.ee_errno = ENOMSG;
1053 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING;
1054 serr->ee.ee_info = SCM_TSTAMP_SCHED;
1055 state = "TX SCH";
1056 break;
1057 case J1939_ERRQUEUE_TX_ABORT:
1058 serr->ee.ee_errno = session->err;
1059 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1060 serr->ee.ee_info = J1939_EE_INFO_TX_ABORT;
1061 state = "TX ABT";
1062 break;
1063 case J1939_ERRQUEUE_RX_RTS:
1064 serr->ee.ee_errno = ENOMSG;
1065 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1066 serr->ee.ee_info = J1939_EE_INFO_RX_RTS;
1067 state = "RX RTS";
1068 break;
1069 case J1939_ERRQUEUE_RX_DPO:
1070 serr->ee.ee_errno = ENOMSG;
1071 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1072 serr->ee.ee_info = J1939_EE_INFO_RX_DPO;
1073 state = "RX DPO";
1074 break;
1075 case J1939_ERRQUEUE_RX_ABORT:
1076 serr->ee.ee_errno = session->err;
1077 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1078 serr->ee.ee_info = J1939_EE_INFO_RX_ABORT;
1079 state = "RX ABT";
1080 break;
1081 }
1082
1083 serr->opt_stats = true;
1084 if (tsflags & SOF_TIMESTAMPING_OPT_ID)
1085 serr->ee.ee_data = session->tskey;
1086
1087 netdev_dbg(session->priv->ndev, "%s: 0x%p tskey: %i, state: %s\n",
1088 __func__, session, session->tskey, state);
1089 err = sock_queue_err_skb(sk, skb);
1090
1091 if (err)
1092 kfree_skb(skb);
1093};
1094
1095void j1939_sk_errqueue(struct j1939_session *session,
1096 enum j1939_sk_errqueue_type type)
1097{
1098 struct j1939_priv *priv = session->priv;
1099 struct j1939_sock *jsk;
1100
1101 if (session->sk) {
1102 /* send TX notifications to the socket of origin */
1103 __j1939_sk_errqueue(session, session->sk, type);
1104 return;
1105 }
1106
1107 /* spread RX notifications to all sockets subscribed to this session */
1108 read_lock_bh(&priv->j1939_socks_lock);
1109 list_for_each_entry(jsk, &priv->j1939_socks, list) {
1110 if (j1939_sk_recv_match_one(jsk, &session->skcb))
1111 __j1939_sk_errqueue(session, &jsk->sk, type);
1112 }
1113 read_unlock_bh(&priv->j1939_socks_lock);
1114};
1115
1116void j1939_sk_send_loop_abort(struct sock *sk, int err)
1117{
1118 struct j1939_sock *jsk = j1939_sk(sk);
1119
1120 if (jsk->state & J1939_SOCK_ERRQUEUE)
1121 return;
1122
1123 sk->sk_err = err;
1124
1125 sk_error_report(sk);
1126}
1127
1128static int j1939_sk_send_loop(struct j1939_priv *priv, struct sock *sk,
1129 struct msghdr *msg, size_t size)
1130
1131{
1132 struct j1939_sock *jsk = j1939_sk(sk);
1133 struct j1939_session *session = j1939_sk_get_incomplete_session(jsk);
1134 struct sk_buff *skb;
1135 size_t segment_size, todo_size;
1136 int ret = 0;
1137
1138 if (session &&
1139 session->total_message_size != session->total_queued_size + size) {
1140 j1939_session_put(session);
1141 return -EIO;
1142 }
1143
1144 todo_size = size;
1145
1146 do {
1147 struct j1939_sk_buff_cb *skcb;
1148
1149 segment_size = min_t(size_t, J1939_MAX_TP_PACKET_SIZE,
1150 todo_size);
1151
1152 /* Allocate skb for one segment */
1153 skb = j1939_sk_alloc_skb(priv->ndev, sk, msg, segment_size,
1154 &ret);
1155 if (ret)
1156 break;
1157
1158 skcb = j1939_skb_to_cb(skb);
1159
1160 if (!session) {
1161 /* at this point the size should be full size
1162 * of the session
1163 */
1164 skcb->offset = 0;
1165 session = j1939_tp_send(priv, skb, size);
1166 if (IS_ERR(session)) {
1167 ret = PTR_ERR(session);
1168 goto kfree_skb;
1169 }
1170 if (j1939_sk_queue_session(session)) {
1171 /* try to activate session if we a
1172 * fist in the queue
1173 */
1174 if (!j1939_session_activate(session)) {
1175 j1939_tp_schedule_txtimer(session, 0);
1176 } else {
1177 ret = -EBUSY;
1178 session->err = ret;
1179 j1939_sk_queue_drop_all(priv, jsk,
1180 EBUSY);
1181 break;
1182 }
1183 }
1184 } else {
1185 skcb->offset = session->total_queued_size;
1186 j1939_session_skb_queue(session, skb);
1187 }
1188
1189 todo_size -= segment_size;
1190 session->total_queued_size += segment_size;
1191 } while (todo_size);
1192
1193 switch (ret) {
1194 case 0: /* OK */
1195 if (todo_size)
1196 netdev_warn(priv->ndev,
1197 "no error found and not completely queued?! %zu\n",
1198 todo_size);
1199 ret = size;
1200 break;
1201 case -ERESTARTSYS:
1202 ret = -EINTR;
1203 fallthrough;
1204 case -EAGAIN: /* OK */
1205 if (todo_size != size)
1206 ret = size - todo_size;
1207 break;
1208 default: /* ERROR */
1209 break;
1210 }
1211
1212 if (session)
1213 j1939_session_put(session);
1214
1215 return ret;
1216
1217 kfree_skb:
1218 kfree_skb(skb);
1219 return ret;
1220}
1221
1222static int j1939_sk_sendmsg(struct socket *sock, struct msghdr *msg,
1223 size_t size)
1224{
1225 struct sock *sk = sock->sk;
1226 struct j1939_sock *jsk = j1939_sk(sk);
1227 struct j1939_priv *priv;
1228 int ifindex;
1229 int ret;
1230
1231 lock_sock(sock->sk);
1232 /* various socket state tests */
1233 if (!(jsk->state & J1939_SOCK_BOUND)) {
1234 ret = -EBADFD;
1235 goto sendmsg_done;
1236 }
1237
1238 priv = jsk->priv;
1239 ifindex = jsk->ifindex;
1240
1241 if (!jsk->addr.src_name && jsk->addr.sa == J1939_NO_ADDR) {
1242 /* no source address assigned yet */
1243 ret = -EBADFD;
1244 goto sendmsg_done;
1245 }
1246
1247 /* deal with provided destination address info */
1248 if (msg->msg_name) {
1249 struct sockaddr_can *addr = msg->msg_name;
1250
1251 if (msg->msg_namelen < J1939_MIN_NAMELEN) {
1252 ret = -EINVAL;
1253 goto sendmsg_done;
1254 }
1255
1256 if (addr->can_family != AF_CAN) {
1257 ret = -EINVAL;
1258 goto sendmsg_done;
1259 }
1260
1261 if (addr->can_ifindex && addr->can_ifindex != ifindex) {
1262 ret = -EBADFD;
1263 goto sendmsg_done;
1264 }
1265
1266 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) &&
1267 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn)) {
1268 ret = -EINVAL;
1269 goto sendmsg_done;
1270 }
1271
1272 if (!addr->can_addr.j1939.name &&
1273 addr->can_addr.j1939.addr == J1939_NO_ADDR &&
1274 !sock_flag(sk, SOCK_BROADCAST)) {
1275 /* broadcast, but SO_BROADCAST not set */
1276 ret = -EACCES;
1277 goto sendmsg_done;
1278 }
1279 } else {
1280 if (!jsk->addr.dst_name && jsk->addr.da == J1939_NO_ADDR &&
1281 !sock_flag(sk, SOCK_BROADCAST)) {
1282 /* broadcast, but SO_BROADCAST not set */
1283 ret = -EACCES;
1284 goto sendmsg_done;
1285 }
1286 }
1287
1288 ret = j1939_sk_send_loop(priv, sk, msg, size);
1289
1290sendmsg_done:
1291 release_sock(sock->sk);
1292
1293 return ret;
1294}
1295
1296void j1939_sk_netdev_event_netdown(struct j1939_priv *priv)
1297{
1298 struct j1939_sock *jsk;
1299 int error_code = ENETDOWN;
1300
1301 read_lock_bh(&priv->j1939_socks_lock);
1302 list_for_each_entry(jsk, &priv->j1939_socks, list) {
1303 jsk->sk.sk_err = error_code;
1304 if (!sock_flag(&jsk->sk, SOCK_DEAD))
1305 sk_error_report(&jsk->sk);
1306
1307 j1939_sk_queue_drop_all(priv, jsk, error_code);
1308 }
1309 read_unlock_bh(&priv->j1939_socks_lock);
1310}
1311
1312void j1939_sk_netdev_event_unregister(struct j1939_priv *priv)
1313{
1314 struct sock *sk;
1315 struct j1939_sock *jsk;
1316 bool wait_rcu = false;
1317
1318rescan: /* The caller is holding a ref on this "priv" via j1939_priv_get_by_ndev(). */
1319 read_lock_bh(&priv->j1939_socks_lock);
1320 list_for_each_entry(jsk, &priv->j1939_socks, list) {
1321 /* Skip if j1939_jsk_add() is not called on this socket. */
1322 if (!(jsk->state & J1939_SOCK_BOUND))
1323 continue;
1324 sk = &jsk->sk;
1325 sock_hold(sk);
1326 read_unlock_bh(&priv->j1939_socks_lock);
1327 /* Check if j1939_jsk_del() is not yet called on this socket after holding
1328 * socket's lock, for both j1939_sk_bind() and j1939_sk_release() call
1329 * j1939_jsk_del() with socket's lock held.
1330 */
1331 lock_sock(sk);
1332 if (jsk->state & J1939_SOCK_BOUND) {
1333 /* Neither j1939_sk_bind() nor j1939_sk_release() called j1939_jsk_del().
1334 * Make this socket no longer bound, by pretending as if j1939_sk_bind()
1335 * dropped old references but did not get new references.
1336 */
1337 j1939_jsk_del(priv, jsk);
1338 j1939_local_ecu_put(priv, jsk->addr.src_name, jsk->addr.sa);
1339 j1939_netdev_stop(priv);
1340 /* Call j1939_priv_put() now and prevent j1939_sk_sock_destruct() from
1341 * calling the corresponding j1939_priv_put().
1342 *
1343 * j1939_sk_sock_destruct() is supposed to call j1939_priv_put() after
1344 * an RCU grace period. But since the caller is holding a ref on this
1345 * "priv", we can defer synchronize_rcu() until immediately before
1346 * the caller calls j1939_priv_put().
1347 */
1348 j1939_priv_put(priv);
1349 jsk->priv = NULL;
1350 wait_rcu = true;
1351 }
1352 release_sock(sk);
1353 sock_put(sk);
1354 goto rescan;
1355 }
1356 read_unlock_bh(&priv->j1939_socks_lock);
1357 if (wait_rcu)
1358 synchronize_rcu();
1359}
1360
1361static int j1939_sk_no_ioctlcmd(struct socket *sock, unsigned int cmd,
1362 unsigned long arg)
1363{
1364 /* no ioctls for socket layer -> hand it down to NIC layer */
1365 return -ENOIOCTLCMD;
1366}
1367
1368static const struct proto_ops j1939_ops = {
1369 .family = PF_CAN,
1370 .release = j1939_sk_release,
1371 .bind = j1939_sk_bind,
1372 .connect = j1939_sk_connect,
1373 .socketpair = sock_no_socketpair,
1374 .accept = sock_no_accept,
1375 .getname = j1939_sk_getname,
1376 .poll = datagram_poll,
1377 .ioctl = j1939_sk_no_ioctlcmd,
1378 .listen = sock_no_listen,
1379 .shutdown = sock_no_shutdown,
1380 .setsockopt = j1939_sk_setsockopt,
1381 .getsockopt = j1939_sk_getsockopt,
1382 .sendmsg = j1939_sk_sendmsg,
1383 .recvmsg = j1939_sk_recvmsg,
1384 .mmap = sock_no_mmap,
1385};
1386
1387static struct proto j1939_proto __read_mostly = {
1388 .name = "CAN_J1939",
1389 .owner = THIS_MODULE,
1390 .obj_size = sizeof(struct j1939_sock),
1391 .init = j1939_sk_init,
1392};
1393
1394const struct can_proto j1939_can_proto = {
1395 .type = SOCK_DGRAM,
1396 .protocol = CAN_J1939,
1397 .ops = &j1939_ops,
1398 .prot = &j1939_proto,
1399};