Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v3.2-rc4 2422 lines 55 kB view raw
1 2/* 3 * DECnet An implementation of the DECnet protocol suite for the LINUX 4 * operating system. DECnet is implemented using the BSD Socket 5 * interface as the means of communication with the user level. 6 * 7 * DECnet Socket Layer Interface 8 * 9 * Authors: Eduardo Marcelo Serrat <emserrat@geocities.com> 10 * Patrick Caulfield <patrick@pandh.demon.co.uk> 11 * 12 * Changes: 13 * Steve Whitehouse: Copied from Eduardo Serrat and Patrick Caulfield's 14 * version of the code. Original copyright preserved 15 * below. 16 * Steve Whitehouse: Some bug fixes, cleaning up some code to make it 17 * compatible with my routing layer. 18 * Steve Whitehouse: Merging changes from Eduardo Serrat and Patrick 19 * Caulfield. 20 * Steve Whitehouse: Further bug fixes, checking module code still works 21 * with new routing layer. 22 * Steve Whitehouse: Additional set/get_sockopt() calls. 23 * Steve Whitehouse: Fixed TIOCINQ ioctl to be same as Eduardo's new 24 * code. 25 * Steve Whitehouse: recvmsg() changed to try and behave in a POSIX like 26 * way. Didn't manage it entirely, but its better. 27 * Steve Whitehouse: ditto for sendmsg(). 28 * Steve Whitehouse: A selection of bug fixes to various things. 29 * Steve Whitehouse: Added TIOCOUTQ ioctl. 30 * Steve Whitehouse: Fixes to username2sockaddr & sockaddr2username. 31 * Steve Whitehouse: Fixes to connect() error returns. 32 * Patrick Caulfield: Fixes to delayed acceptance logic. 33 * David S. Miller: New socket locking 34 * Steve Whitehouse: Socket list hashing/locking 35 * Arnaldo C. Melo: use capable, not suser 36 * Steve Whitehouse: Removed unused code. Fix to use sk->allocation 37 * when required. 38 * Patrick Caulfield: /proc/net/decnet now has object name/number 39 * Steve Whitehouse: Fixed local port allocation, hashed sk list 40 * Matthew Wilcox: Fixes for dn_ioctl() 41 * Steve Whitehouse: New connect/accept logic to allow timeouts and 42 * prepare for sendpage etc. 43 */ 44 45 46/****************************************************************************** 47 (c) 1995-1998 E.M. Serrat emserrat@geocities.com 48 49 This program is free software; you can redistribute it and/or modify 50 it under the terms of the GNU General Public License as published by 51 the Free Software Foundation; either version 2 of the License, or 52 any later version. 53 54 This program is distributed in the hope that it will be useful, 55 but WITHOUT ANY WARRANTY; without even the implied warranty of 56 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 57 GNU General Public License for more details. 58 59HISTORY: 60 61Version Kernel Date Author/Comments 62------- ------ ---- --------------- 63Version 0.0.1 2.0.30 01-dic-97 Eduardo Marcelo Serrat 64 (emserrat@geocities.com) 65 66 First Development of DECnet Socket La- 67 yer for Linux. Only supports outgoing 68 connections. 69 70Version 0.0.2 2.1.105 20-jun-98 Patrick J. Caulfield 71 (patrick@pandh.demon.co.uk) 72 73 Port to new kernel development version. 74 75Version 0.0.3 2.1.106 25-jun-98 Eduardo Marcelo Serrat 76 (emserrat@geocities.com) 77 _ 78 Added support for incoming connections 79 so we can start developing server apps 80 on Linux. 81 - 82 Module Support 83Version 0.0.4 2.1.109 21-jul-98 Eduardo Marcelo Serrat 84 (emserrat@geocities.com) 85 _ 86 Added support for X11R6.4. Now we can 87 use DECnet transport for X on Linux!!! 88 - 89Version 0.0.5 2.1.110 01-aug-98 Eduardo Marcelo Serrat 90 (emserrat@geocities.com) 91 Removed bugs on flow control 92 Removed bugs on incoming accessdata 93 order 94 - 95Version 0.0.6 2.1.110 07-aug-98 Eduardo Marcelo Serrat 96 dn_recvmsg fixes 97 98 Patrick J. Caulfield 99 dn_bind fixes 100*******************************************************************************/ 101 102#include <linux/module.h> 103#include <linux/errno.h> 104#include <linux/types.h> 105#include <linux/slab.h> 106#include <linux/socket.h> 107#include <linux/in.h> 108#include <linux/kernel.h> 109#include <linux/sched.h> 110#include <linux/timer.h> 111#include <linux/string.h> 112#include <linux/sockios.h> 113#include <linux/net.h> 114#include <linux/netdevice.h> 115#include <linux/inet.h> 116#include <linux/route.h> 117#include <linux/netfilter.h> 118#include <linux/seq_file.h> 119#include <net/sock.h> 120#include <net/tcp_states.h> 121#include <net/flow.h> 122#include <asm/system.h> 123#include <asm/ioctls.h> 124#include <linux/capability.h> 125#include <linux/mm.h> 126#include <linux/interrupt.h> 127#include <linux/proc_fs.h> 128#include <linux/stat.h> 129#include <linux/init.h> 130#include <linux/poll.h> 131#include <net/net_namespace.h> 132#include <net/neighbour.h> 133#include <net/dst.h> 134#include <net/fib_rules.h> 135#include <net/dn.h> 136#include <net/dn_nsp.h> 137#include <net/dn_dev.h> 138#include <net/dn_route.h> 139#include <net/dn_fib.h> 140#include <net/dn_neigh.h> 141 142struct dn_sock { 143 struct sock sk; 144 struct dn_scp scp; 145}; 146 147static void dn_keepalive(struct sock *sk); 148 149#define DN_SK_HASH_SHIFT 8 150#define DN_SK_HASH_SIZE (1 << DN_SK_HASH_SHIFT) 151#define DN_SK_HASH_MASK (DN_SK_HASH_SIZE - 1) 152 153 154static const struct proto_ops dn_proto_ops; 155static DEFINE_RWLOCK(dn_hash_lock); 156static struct hlist_head dn_sk_hash[DN_SK_HASH_SIZE]; 157static struct hlist_head dn_wild_sk; 158static atomic_long_t decnet_memory_allocated; 159 160static int __dn_setsockopt(struct socket *sock, int level, int optname, char __user *optval, unsigned int optlen, int flags); 161static int __dn_getsockopt(struct socket *sock, int level, int optname, char __user *optval, int __user *optlen, int flags); 162 163static struct hlist_head *dn_find_list(struct sock *sk) 164{ 165 struct dn_scp *scp = DN_SK(sk); 166 167 if (scp->addr.sdn_flags & SDF_WILD) 168 return hlist_empty(&dn_wild_sk) ? &dn_wild_sk : NULL; 169 170 return &dn_sk_hash[le16_to_cpu(scp->addrloc) & DN_SK_HASH_MASK]; 171} 172 173/* 174 * Valid ports are those greater than zero and not already in use. 175 */ 176static int check_port(__le16 port) 177{ 178 struct sock *sk; 179 struct hlist_node *node; 180 181 if (port == 0) 182 return -1; 183 184 sk_for_each(sk, node, &dn_sk_hash[le16_to_cpu(port) & DN_SK_HASH_MASK]) { 185 struct dn_scp *scp = DN_SK(sk); 186 if (scp->addrloc == port) 187 return -1; 188 } 189 return 0; 190} 191 192static unsigned short port_alloc(struct sock *sk) 193{ 194 struct dn_scp *scp = DN_SK(sk); 195static unsigned short port = 0x2000; 196 unsigned short i_port = port; 197 198 while(check_port(cpu_to_le16(++port)) != 0) { 199 if (port == i_port) 200 return 0; 201 } 202 203 scp->addrloc = cpu_to_le16(port); 204 205 return 1; 206} 207 208/* 209 * Since this is only ever called from user 210 * level, we don't need a write_lock() version 211 * of this. 212 */ 213static int dn_hash_sock(struct sock *sk) 214{ 215 struct dn_scp *scp = DN_SK(sk); 216 struct hlist_head *list; 217 int rv = -EUSERS; 218 219 BUG_ON(sk_hashed(sk)); 220 221 write_lock_bh(&dn_hash_lock); 222 223 if (!scp->addrloc && !port_alloc(sk)) 224 goto out; 225 226 rv = -EADDRINUSE; 227 if ((list = dn_find_list(sk)) == NULL) 228 goto out; 229 230 sk_add_node(sk, list); 231 rv = 0; 232out: 233 write_unlock_bh(&dn_hash_lock); 234 return rv; 235} 236 237static void dn_unhash_sock(struct sock *sk) 238{ 239 write_lock(&dn_hash_lock); 240 sk_del_node_init(sk); 241 write_unlock(&dn_hash_lock); 242} 243 244static void dn_unhash_sock_bh(struct sock *sk) 245{ 246 write_lock_bh(&dn_hash_lock); 247 sk_del_node_init(sk); 248 write_unlock_bh(&dn_hash_lock); 249} 250 251static struct hlist_head *listen_hash(struct sockaddr_dn *addr) 252{ 253 int i; 254 unsigned hash = addr->sdn_objnum; 255 256 if (hash == 0) { 257 hash = addr->sdn_objnamel; 258 for(i = 0; i < le16_to_cpu(addr->sdn_objnamel); i++) { 259 hash ^= addr->sdn_objname[i]; 260 hash ^= (hash << 3); 261 } 262 } 263 264 return &dn_sk_hash[hash & DN_SK_HASH_MASK]; 265} 266 267/* 268 * Called to transform a socket from bound (i.e. with a local address) 269 * into a listening socket (doesn't need a local port number) and rehashes 270 * based upon the object name/number. 271 */ 272static void dn_rehash_sock(struct sock *sk) 273{ 274 struct hlist_head *list; 275 struct dn_scp *scp = DN_SK(sk); 276 277 if (scp->addr.sdn_flags & SDF_WILD) 278 return; 279 280 write_lock_bh(&dn_hash_lock); 281 sk_del_node_init(sk); 282 DN_SK(sk)->addrloc = 0; 283 list = listen_hash(&DN_SK(sk)->addr); 284 sk_add_node(sk, list); 285 write_unlock_bh(&dn_hash_lock); 286} 287 288int dn_sockaddr2username(struct sockaddr_dn *sdn, unsigned char *buf, unsigned char type) 289{ 290 int len = 2; 291 292 *buf++ = type; 293 294 switch (type) { 295 case 0: 296 *buf++ = sdn->sdn_objnum; 297 break; 298 case 1: 299 *buf++ = 0; 300 *buf++ = le16_to_cpu(sdn->sdn_objnamel); 301 memcpy(buf, sdn->sdn_objname, le16_to_cpu(sdn->sdn_objnamel)); 302 len = 3 + le16_to_cpu(sdn->sdn_objnamel); 303 break; 304 case 2: 305 memset(buf, 0, 5); 306 buf += 5; 307 *buf++ = le16_to_cpu(sdn->sdn_objnamel); 308 memcpy(buf, sdn->sdn_objname, le16_to_cpu(sdn->sdn_objnamel)); 309 len = 7 + le16_to_cpu(sdn->sdn_objnamel); 310 break; 311 } 312 313 return len; 314} 315 316/* 317 * On reception of usernames, we handle types 1 and 0 for destination 318 * addresses only. Types 2 and 4 are used for source addresses, but the 319 * UIC, GIC are ignored and they are both treated the same way. Type 3 320 * is never used as I've no idea what its purpose might be or what its 321 * format is. 322 */ 323int dn_username2sockaddr(unsigned char *data, int len, struct sockaddr_dn *sdn, unsigned char *fmt) 324{ 325 unsigned char type; 326 int size = len; 327 int namel = 12; 328 329 sdn->sdn_objnum = 0; 330 sdn->sdn_objnamel = cpu_to_le16(0); 331 memset(sdn->sdn_objname, 0, DN_MAXOBJL); 332 333 if (len < 2) 334 return -1; 335 336 len -= 2; 337 *fmt = *data++; 338 type = *data++; 339 340 switch (*fmt) { 341 case 0: 342 sdn->sdn_objnum = type; 343 return 2; 344 case 1: 345 namel = 16; 346 break; 347 case 2: 348 len -= 4; 349 data += 4; 350 break; 351 case 4: 352 len -= 8; 353 data += 8; 354 break; 355 default: 356 return -1; 357 } 358 359 len -= 1; 360 361 if (len < 0) 362 return -1; 363 364 sdn->sdn_objnamel = cpu_to_le16(*data++); 365 len -= le16_to_cpu(sdn->sdn_objnamel); 366 367 if ((len < 0) || (le16_to_cpu(sdn->sdn_objnamel) > namel)) 368 return -1; 369 370 memcpy(sdn->sdn_objname, data, le16_to_cpu(sdn->sdn_objnamel)); 371 372 return size - len; 373} 374 375struct sock *dn_sklist_find_listener(struct sockaddr_dn *addr) 376{ 377 struct hlist_head *list = listen_hash(addr); 378 struct hlist_node *node; 379 struct sock *sk; 380 381 read_lock(&dn_hash_lock); 382 sk_for_each(sk, node, list) { 383 struct dn_scp *scp = DN_SK(sk); 384 if (sk->sk_state != TCP_LISTEN) 385 continue; 386 if (scp->addr.sdn_objnum) { 387 if (scp->addr.sdn_objnum != addr->sdn_objnum) 388 continue; 389 } else { 390 if (addr->sdn_objnum) 391 continue; 392 if (scp->addr.sdn_objnamel != addr->sdn_objnamel) 393 continue; 394 if (memcmp(scp->addr.sdn_objname, addr->sdn_objname, le16_to_cpu(addr->sdn_objnamel)) != 0) 395 continue; 396 } 397 sock_hold(sk); 398 read_unlock(&dn_hash_lock); 399 return sk; 400 } 401 402 sk = sk_head(&dn_wild_sk); 403 if (sk) { 404 if (sk->sk_state == TCP_LISTEN) 405 sock_hold(sk); 406 else 407 sk = NULL; 408 } 409 410 read_unlock(&dn_hash_lock); 411 return sk; 412} 413 414struct sock *dn_find_by_skb(struct sk_buff *skb) 415{ 416 struct dn_skb_cb *cb = DN_SKB_CB(skb); 417 struct sock *sk; 418 struct hlist_node *node; 419 struct dn_scp *scp; 420 421 read_lock(&dn_hash_lock); 422 sk_for_each(sk, node, &dn_sk_hash[le16_to_cpu(cb->dst_port) & DN_SK_HASH_MASK]) { 423 scp = DN_SK(sk); 424 if (cb->src != dn_saddr2dn(&scp->peer)) 425 continue; 426 if (cb->dst_port != scp->addrloc) 427 continue; 428 if (scp->addrrem && (cb->src_port != scp->addrrem)) 429 continue; 430 sock_hold(sk); 431 goto found; 432 } 433 sk = NULL; 434found: 435 read_unlock(&dn_hash_lock); 436 return sk; 437} 438 439 440 441static void dn_destruct(struct sock *sk) 442{ 443 struct dn_scp *scp = DN_SK(sk); 444 445 skb_queue_purge(&scp->data_xmit_queue); 446 skb_queue_purge(&scp->other_xmit_queue); 447 skb_queue_purge(&scp->other_receive_queue); 448 449 dst_release(rcu_dereference_check(sk->sk_dst_cache, 1)); 450} 451 452static int dn_memory_pressure; 453 454static void dn_enter_memory_pressure(struct sock *sk) 455{ 456 if (!dn_memory_pressure) { 457 dn_memory_pressure = 1; 458 } 459} 460 461static struct proto dn_proto = { 462 .name = "NSP", 463 .owner = THIS_MODULE, 464 .enter_memory_pressure = dn_enter_memory_pressure, 465 .memory_pressure = &dn_memory_pressure, 466 .memory_allocated = &decnet_memory_allocated, 467 .sysctl_mem = sysctl_decnet_mem, 468 .sysctl_wmem = sysctl_decnet_wmem, 469 .sysctl_rmem = sysctl_decnet_rmem, 470 .max_header = DN_MAX_NSP_DATA_HEADER + 64, 471 .obj_size = sizeof(struct dn_sock), 472}; 473 474static struct sock *dn_alloc_sock(struct net *net, struct socket *sock, gfp_t gfp) 475{ 476 struct dn_scp *scp; 477 struct sock *sk = sk_alloc(net, PF_DECnet, gfp, &dn_proto); 478 479 if (!sk) 480 goto out; 481 482 if (sock) 483 sock->ops = &dn_proto_ops; 484 sock_init_data(sock, sk); 485 486 sk->sk_backlog_rcv = dn_nsp_backlog_rcv; 487 sk->sk_destruct = dn_destruct; 488 sk->sk_no_check = 1; 489 sk->sk_family = PF_DECnet; 490 sk->sk_protocol = 0; 491 sk->sk_allocation = gfp; 492 sk->sk_sndbuf = sysctl_decnet_wmem[1]; 493 sk->sk_rcvbuf = sysctl_decnet_rmem[1]; 494 495 /* Initialization of DECnet Session Control Port */ 496 scp = DN_SK(sk); 497 scp->state = DN_O; /* Open */ 498 scp->numdat = 1; /* Next data seg to tx */ 499 scp->numoth = 1; /* Next oth data to tx */ 500 scp->ackxmt_dat = 0; /* Last data seg ack'ed */ 501 scp->ackxmt_oth = 0; /* Last oth data ack'ed */ 502 scp->ackrcv_dat = 0; /* Highest data ack recv*/ 503 scp->ackrcv_oth = 0; /* Last oth data ack rec*/ 504 scp->flowrem_sw = DN_SEND; 505 scp->flowloc_sw = DN_SEND; 506 scp->flowrem_dat = 0; 507 scp->flowrem_oth = 1; 508 scp->flowloc_dat = 0; 509 scp->flowloc_oth = 1; 510 scp->services_rem = 0; 511 scp->services_loc = 1 | NSP_FC_NONE; 512 scp->info_rem = 0; 513 scp->info_loc = 0x03; /* NSP version 4.1 */ 514 scp->segsize_rem = 230 - DN_MAX_NSP_DATA_HEADER; /* Default: Updated by remote segsize */ 515 scp->nonagle = 0; 516 scp->multi_ireq = 1; 517 scp->accept_mode = ACC_IMMED; 518 scp->addr.sdn_family = AF_DECnet; 519 scp->peer.sdn_family = AF_DECnet; 520 scp->accessdata.acc_accl = 5; 521 memcpy(scp->accessdata.acc_acc, "LINUX", 5); 522 523 scp->max_window = NSP_MAX_WINDOW; 524 scp->snd_window = NSP_MIN_WINDOW; 525 scp->nsp_srtt = NSP_INITIAL_SRTT; 526 scp->nsp_rttvar = NSP_INITIAL_RTTVAR; 527 scp->nsp_rxtshift = 0; 528 529 skb_queue_head_init(&scp->data_xmit_queue); 530 skb_queue_head_init(&scp->other_xmit_queue); 531 skb_queue_head_init(&scp->other_receive_queue); 532 533 scp->persist = 0; 534 scp->persist_fxn = NULL; 535 scp->keepalive = 10 * HZ; 536 scp->keepalive_fxn = dn_keepalive; 537 538 init_timer(&scp->delack_timer); 539 scp->delack_pending = 0; 540 scp->delack_fxn = dn_nsp_delayed_ack; 541 542 dn_start_slow_timer(sk); 543out: 544 return sk; 545} 546 547/* 548 * Keepalive timer. 549 * FIXME: Should respond to SO_KEEPALIVE etc. 550 */ 551static void dn_keepalive(struct sock *sk) 552{ 553 struct dn_scp *scp = DN_SK(sk); 554 555 /* 556 * By checking the other_data transmit queue is empty 557 * we are double checking that we are not sending too 558 * many of these keepalive frames. 559 */ 560 if (skb_queue_empty(&scp->other_xmit_queue)) 561 dn_nsp_send_link(sk, DN_NOCHANGE, 0); 562} 563 564 565/* 566 * Timer for shutdown/destroyed sockets. 567 * When socket is dead & no packets have been sent for a 568 * certain amount of time, they are removed by this 569 * routine. Also takes care of sending out DI & DC 570 * frames at correct times. 571 */ 572int dn_destroy_timer(struct sock *sk) 573{ 574 struct dn_scp *scp = DN_SK(sk); 575 576 scp->persist = dn_nsp_persist(sk); 577 578 switch (scp->state) { 579 case DN_DI: 580 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, GFP_ATOMIC); 581 if (scp->nsp_rxtshift >= decnet_di_count) 582 scp->state = DN_CN; 583 return 0; 584 585 case DN_DR: 586 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, GFP_ATOMIC); 587 if (scp->nsp_rxtshift >= decnet_dr_count) 588 scp->state = DN_DRC; 589 return 0; 590 591 case DN_DN: 592 if (scp->nsp_rxtshift < decnet_dn_count) { 593 /* printk(KERN_DEBUG "dn_destroy_timer: DN\n"); */ 594 dn_nsp_send_disc(sk, NSP_DISCCONF, NSP_REASON_DC, 595 GFP_ATOMIC); 596 return 0; 597 } 598 } 599 600 scp->persist = (HZ * decnet_time_wait); 601 602 if (sk->sk_socket) 603 return 0; 604 605 if ((jiffies - scp->stamp) >= (HZ * decnet_time_wait)) { 606 dn_unhash_sock(sk); 607 sock_put(sk); 608 return 1; 609 } 610 611 return 0; 612} 613 614static void dn_destroy_sock(struct sock *sk) 615{ 616 struct dn_scp *scp = DN_SK(sk); 617 618 scp->nsp_rxtshift = 0; /* reset back off */ 619 620 if (sk->sk_socket) { 621 if (sk->sk_socket->state != SS_UNCONNECTED) 622 sk->sk_socket->state = SS_DISCONNECTING; 623 } 624 625 sk->sk_state = TCP_CLOSE; 626 627 switch (scp->state) { 628 case DN_DN: 629 dn_nsp_send_disc(sk, NSP_DISCCONF, NSP_REASON_DC, 630 sk->sk_allocation); 631 scp->persist_fxn = dn_destroy_timer; 632 scp->persist = dn_nsp_persist(sk); 633 break; 634 case DN_CR: 635 scp->state = DN_DR; 636 goto disc_reject; 637 case DN_RUN: 638 scp->state = DN_DI; 639 case DN_DI: 640 case DN_DR: 641disc_reject: 642 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, sk->sk_allocation); 643 case DN_NC: 644 case DN_NR: 645 case DN_RJ: 646 case DN_DIC: 647 case DN_CN: 648 case DN_DRC: 649 case DN_CI: 650 case DN_CD: 651 scp->persist_fxn = dn_destroy_timer; 652 scp->persist = dn_nsp_persist(sk); 653 break; 654 default: 655 printk(KERN_DEBUG "DECnet: dn_destroy_sock passed socket in invalid state\n"); 656 case DN_O: 657 dn_stop_slow_timer(sk); 658 659 dn_unhash_sock_bh(sk); 660 sock_put(sk); 661 662 break; 663 } 664} 665 666char *dn_addr2asc(__u16 addr, char *buf) 667{ 668 unsigned short node, area; 669 670 node = addr & 0x03ff; 671 area = addr >> 10; 672 sprintf(buf, "%hd.%hd", area, node); 673 674 return buf; 675} 676 677 678 679static int dn_create(struct net *net, struct socket *sock, int protocol, 680 int kern) 681{ 682 struct sock *sk; 683 684 if (!net_eq(net, &init_net)) 685 return -EAFNOSUPPORT; 686 687 switch (sock->type) { 688 case SOCK_SEQPACKET: 689 if (protocol != DNPROTO_NSP) 690 return -EPROTONOSUPPORT; 691 break; 692 case SOCK_STREAM: 693 break; 694 default: 695 return -ESOCKTNOSUPPORT; 696 } 697 698 699 if ((sk = dn_alloc_sock(net, sock, GFP_KERNEL)) == NULL) 700 return -ENOBUFS; 701 702 sk->sk_protocol = protocol; 703 704 return 0; 705} 706 707 708static int 709dn_release(struct socket *sock) 710{ 711 struct sock *sk = sock->sk; 712 713 if (sk) { 714 sock_orphan(sk); 715 sock_hold(sk); 716 lock_sock(sk); 717 dn_destroy_sock(sk); 718 release_sock(sk); 719 sock_put(sk); 720 } 721 722 return 0; 723} 724 725static int dn_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) 726{ 727 struct sock *sk = sock->sk; 728 struct dn_scp *scp = DN_SK(sk); 729 struct sockaddr_dn *saddr = (struct sockaddr_dn *)uaddr; 730 struct net_device *dev, *ldev; 731 int rv; 732 733 if (addr_len != sizeof(struct sockaddr_dn)) 734 return -EINVAL; 735 736 if (saddr->sdn_family != AF_DECnet) 737 return -EINVAL; 738 739 if (le16_to_cpu(saddr->sdn_nodeaddrl) && (le16_to_cpu(saddr->sdn_nodeaddrl) != 2)) 740 return -EINVAL; 741 742 if (le16_to_cpu(saddr->sdn_objnamel) > DN_MAXOBJL) 743 return -EINVAL; 744 745 if (saddr->sdn_flags & ~SDF_WILD) 746 return -EINVAL; 747 748 if (!capable(CAP_NET_BIND_SERVICE) && (saddr->sdn_objnum || 749 (saddr->sdn_flags & SDF_WILD))) 750 return -EACCES; 751 752 if (!(saddr->sdn_flags & SDF_WILD)) { 753 if (le16_to_cpu(saddr->sdn_nodeaddrl)) { 754 rcu_read_lock(); 755 ldev = NULL; 756 for_each_netdev_rcu(&init_net, dev) { 757 if (!dev->dn_ptr) 758 continue; 759 if (dn_dev_islocal(dev, dn_saddr2dn(saddr))) { 760 ldev = dev; 761 break; 762 } 763 } 764 rcu_read_unlock(); 765 if (ldev == NULL) 766 return -EADDRNOTAVAIL; 767 } 768 } 769 770 rv = -EINVAL; 771 lock_sock(sk); 772 if (sock_flag(sk, SOCK_ZAPPED)) { 773 memcpy(&scp->addr, saddr, addr_len); 774 sock_reset_flag(sk, SOCK_ZAPPED); 775 776 rv = dn_hash_sock(sk); 777 if (rv) 778 sock_set_flag(sk, SOCK_ZAPPED); 779 } 780 release_sock(sk); 781 782 return rv; 783} 784 785 786static int dn_auto_bind(struct socket *sock) 787{ 788 struct sock *sk = sock->sk; 789 struct dn_scp *scp = DN_SK(sk); 790 int rv; 791 792 sock_reset_flag(sk, SOCK_ZAPPED); 793 794 scp->addr.sdn_flags = 0; 795 scp->addr.sdn_objnum = 0; 796 797 /* 798 * This stuff is to keep compatibility with Eduardo's 799 * patch. I hope I can dispense with it shortly... 800 */ 801 if ((scp->accessdata.acc_accl != 0) && 802 (scp->accessdata.acc_accl <= 12)) { 803 804 scp->addr.sdn_objnamel = cpu_to_le16(scp->accessdata.acc_accl); 805 memcpy(scp->addr.sdn_objname, scp->accessdata.acc_acc, le16_to_cpu(scp->addr.sdn_objnamel)); 806 807 scp->accessdata.acc_accl = 0; 808 memset(scp->accessdata.acc_acc, 0, 40); 809 } 810 /* End of compatibility stuff */ 811 812 scp->addr.sdn_add.a_len = cpu_to_le16(2); 813 rv = dn_dev_bind_default((__le16 *)scp->addr.sdn_add.a_addr); 814 if (rv == 0) { 815 rv = dn_hash_sock(sk); 816 if (rv) 817 sock_set_flag(sk, SOCK_ZAPPED); 818 } 819 820 return rv; 821} 822 823static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation) 824{ 825 struct dn_scp *scp = DN_SK(sk); 826 DEFINE_WAIT(wait); 827 int err; 828 829 if (scp->state != DN_CR) 830 return -EINVAL; 831 832 scp->state = DN_CC; 833 scp->segsize_loc = dst_metric_advmss(__sk_dst_get(sk)); 834 dn_send_conn_conf(sk, allocation); 835 836 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 837 for(;;) { 838 release_sock(sk); 839 if (scp->state == DN_CC) 840 *timeo = schedule_timeout(*timeo); 841 lock_sock(sk); 842 err = 0; 843 if (scp->state == DN_RUN) 844 break; 845 err = sock_error(sk); 846 if (err) 847 break; 848 err = sock_intr_errno(*timeo); 849 if (signal_pending(current)) 850 break; 851 err = -EAGAIN; 852 if (!*timeo) 853 break; 854 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 855 } 856 finish_wait(sk_sleep(sk), &wait); 857 if (err == 0) { 858 sk->sk_socket->state = SS_CONNECTED; 859 } else if (scp->state != DN_CC) { 860 sk->sk_socket->state = SS_UNCONNECTED; 861 } 862 return err; 863} 864 865static int dn_wait_run(struct sock *sk, long *timeo) 866{ 867 struct dn_scp *scp = DN_SK(sk); 868 DEFINE_WAIT(wait); 869 int err = 0; 870 871 if (scp->state == DN_RUN) 872 goto out; 873 874 if (!*timeo) 875 return -EALREADY; 876 877 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 878 for(;;) { 879 release_sock(sk); 880 if (scp->state == DN_CI || scp->state == DN_CC) 881 *timeo = schedule_timeout(*timeo); 882 lock_sock(sk); 883 err = 0; 884 if (scp->state == DN_RUN) 885 break; 886 err = sock_error(sk); 887 if (err) 888 break; 889 err = sock_intr_errno(*timeo); 890 if (signal_pending(current)) 891 break; 892 err = -ETIMEDOUT; 893 if (!*timeo) 894 break; 895 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 896 } 897 finish_wait(sk_sleep(sk), &wait); 898out: 899 if (err == 0) { 900 sk->sk_socket->state = SS_CONNECTED; 901 } else if (scp->state != DN_CI && scp->state != DN_CC) { 902 sk->sk_socket->state = SS_UNCONNECTED; 903 } 904 return err; 905} 906 907static int __dn_connect(struct sock *sk, struct sockaddr_dn *addr, int addrlen, long *timeo, int flags) 908{ 909 struct socket *sock = sk->sk_socket; 910 struct dn_scp *scp = DN_SK(sk); 911 int err = -EISCONN; 912 struct flowidn fld; 913 914 if (sock->state == SS_CONNECTED) 915 goto out; 916 917 if (sock->state == SS_CONNECTING) { 918 err = 0; 919 if (scp->state == DN_RUN) { 920 sock->state = SS_CONNECTED; 921 goto out; 922 } 923 err = -ECONNREFUSED; 924 if (scp->state != DN_CI && scp->state != DN_CC) { 925 sock->state = SS_UNCONNECTED; 926 goto out; 927 } 928 return dn_wait_run(sk, timeo); 929 } 930 931 err = -EINVAL; 932 if (scp->state != DN_O) 933 goto out; 934 935 if (addr == NULL || addrlen != sizeof(struct sockaddr_dn)) 936 goto out; 937 if (addr->sdn_family != AF_DECnet) 938 goto out; 939 if (addr->sdn_flags & SDF_WILD) 940 goto out; 941 942 if (sock_flag(sk, SOCK_ZAPPED)) { 943 err = dn_auto_bind(sk->sk_socket); 944 if (err) 945 goto out; 946 } 947 948 memcpy(&scp->peer, addr, sizeof(struct sockaddr_dn)); 949 950 err = -EHOSTUNREACH; 951 memset(&fld, 0, sizeof(fld)); 952 fld.flowidn_oif = sk->sk_bound_dev_if; 953 fld.daddr = dn_saddr2dn(&scp->peer); 954 fld.saddr = dn_saddr2dn(&scp->addr); 955 dn_sk_ports_copy(&fld, scp); 956 fld.flowidn_proto = DNPROTO_NSP; 957 if (dn_route_output_sock(&sk->sk_dst_cache, &fld, sk, flags) < 0) 958 goto out; 959 sk->sk_route_caps = sk->sk_dst_cache->dev->features; 960 sock->state = SS_CONNECTING; 961 scp->state = DN_CI; 962 scp->segsize_loc = dst_metric_advmss(sk->sk_dst_cache); 963 964 dn_nsp_send_conninit(sk, NSP_CI); 965 err = -EINPROGRESS; 966 if (*timeo) { 967 err = dn_wait_run(sk, timeo); 968 } 969out: 970 return err; 971} 972 973static int dn_connect(struct socket *sock, struct sockaddr *uaddr, int addrlen, int flags) 974{ 975 struct sockaddr_dn *addr = (struct sockaddr_dn *)uaddr; 976 struct sock *sk = sock->sk; 977 int err; 978 long timeo = sock_sndtimeo(sk, flags & O_NONBLOCK); 979 980 lock_sock(sk); 981 err = __dn_connect(sk, addr, addrlen, &timeo, 0); 982 release_sock(sk); 983 984 return err; 985} 986 987static inline int dn_check_state(struct sock *sk, struct sockaddr_dn *addr, int addrlen, long *timeo, int flags) 988{ 989 struct dn_scp *scp = DN_SK(sk); 990 991 switch (scp->state) { 992 case DN_RUN: 993 return 0; 994 case DN_CR: 995 return dn_confirm_accept(sk, timeo, sk->sk_allocation); 996 case DN_CI: 997 case DN_CC: 998 return dn_wait_run(sk, timeo); 999 case DN_O: 1000 return __dn_connect(sk, addr, addrlen, timeo, flags); 1001 } 1002 1003 return -EINVAL; 1004} 1005 1006 1007static void dn_access_copy(struct sk_buff *skb, struct accessdata_dn *acc) 1008{ 1009 unsigned char *ptr = skb->data; 1010 1011 acc->acc_userl = *ptr++; 1012 memcpy(&acc->acc_user, ptr, acc->acc_userl); 1013 ptr += acc->acc_userl; 1014 1015 acc->acc_passl = *ptr++; 1016 memcpy(&acc->acc_pass, ptr, acc->acc_passl); 1017 ptr += acc->acc_passl; 1018 1019 acc->acc_accl = *ptr++; 1020 memcpy(&acc->acc_acc, ptr, acc->acc_accl); 1021 1022 skb_pull(skb, acc->acc_accl + acc->acc_passl + acc->acc_userl + 3); 1023 1024} 1025 1026static void dn_user_copy(struct sk_buff *skb, struct optdata_dn *opt) 1027{ 1028 unsigned char *ptr = skb->data; 1029 u16 len = *ptr++; /* yes, it's 8bit on the wire */ 1030 1031 BUG_ON(len > 16); /* we've checked the contents earlier */ 1032 opt->opt_optl = cpu_to_le16(len); 1033 opt->opt_status = 0; 1034 memcpy(opt->opt_data, ptr, len); 1035 skb_pull(skb, len + 1); 1036} 1037 1038static struct sk_buff *dn_wait_for_connect(struct sock *sk, long *timeo) 1039{ 1040 DEFINE_WAIT(wait); 1041 struct sk_buff *skb = NULL; 1042 int err = 0; 1043 1044 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 1045 for(;;) { 1046 release_sock(sk); 1047 skb = skb_dequeue(&sk->sk_receive_queue); 1048 if (skb == NULL) { 1049 *timeo = schedule_timeout(*timeo); 1050 skb = skb_dequeue(&sk->sk_receive_queue); 1051 } 1052 lock_sock(sk); 1053 if (skb != NULL) 1054 break; 1055 err = -EINVAL; 1056 if (sk->sk_state != TCP_LISTEN) 1057 break; 1058 err = sock_intr_errno(*timeo); 1059 if (signal_pending(current)) 1060 break; 1061 err = -EAGAIN; 1062 if (!*timeo) 1063 break; 1064 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 1065 } 1066 finish_wait(sk_sleep(sk), &wait); 1067 1068 return skb == NULL ? ERR_PTR(err) : skb; 1069} 1070 1071static int dn_accept(struct socket *sock, struct socket *newsock, int flags) 1072{ 1073 struct sock *sk = sock->sk, *newsk; 1074 struct sk_buff *skb = NULL; 1075 struct dn_skb_cb *cb; 1076 unsigned char menuver; 1077 int err = 0; 1078 unsigned char type; 1079 long timeo = sock_rcvtimeo(sk, flags & O_NONBLOCK); 1080 struct dst_entry *dst; 1081 1082 lock_sock(sk); 1083 1084 if (sk->sk_state != TCP_LISTEN || DN_SK(sk)->state != DN_O) { 1085 release_sock(sk); 1086 return -EINVAL; 1087 } 1088 1089 skb = skb_dequeue(&sk->sk_receive_queue); 1090 if (skb == NULL) { 1091 skb = dn_wait_for_connect(sk, &timeo); 1092 if (IS_ERR(skb)) { 1093 release_sock(sk); 1094 return PTR_ERR(skb); 1095 } 1096 } 1097 1098 cb = DN_SKB_CB(skb); 1099 sk->sk_ack_backlog--; 1100 newsk = dn_alloc_sock(sock_net(sk), newsock, sk->sk_allocation); 1101 if (newsk == NULL) { 1102 release_sock(sk); 1103 kfree_skb(skb); 1104 return -ENOBUFS; 1105 } 1106 release_sock(sk); 1107 1108 dst = skb_dst(skb); 1109 sk_dst_set(newsk, dst); 1110 skb_dst_set(skb, NULL); 1111 1112 DN_SK(newsk)->state = DN_CR; 1113 DN_SK(newsk)->addrrem = cb->src_port; 1114 DN_SK(newsk)->services_rem = cb->services; 1115 DN_SK(newsk)->info_rem = cb->info; 1116 DN_SK(newsk)->segsize_rem = cb->segsize; 1117 DN_SK(newsk)->accept_mode = DN_SK(sk)->accept_mode; 1118 1119 if (DN_SK(newsk)->segsize_rem < 230) 1120 DN_SK(newsk)->segsize_rem = 230; 1121 1122 if ((DN_SK(newsk)->services_rem & NSP_FC_MASK) == NSP_FC_NONE) 1123 DN_SK(newsk)->max_window = decnet_no_fc_max_cwnd; 1124 1125 newsk->sk_state = TCP_LISTEN; 1126 memcpy(&(DN_SK(newsk)->addr), &(DN_SK(sk)->addr), sizeof(struct sockaddr_dn)); 1127 1128 /* 1129 * If we are listening on a wild socket, we don't want 1130 * the newly created socket on the wrong hash queue. 1131 */ 1132 DN_SK(newsk)->addr.sdn_flags &= ~SDF_WILD; 1133 1134 skb_pull(skb, dn_username2sockaddr(skb->data, skb->len, &(DN_SK(newsk)->addr), &type)); 1135 skb_pull(skb, dn_username2sockaddr(skb->data, skb->len, &(DN_SK(newsk)->peer), &type)); 1136 *(__le16 *)(DN_SK(newsk)->peer.sdn_add.a_addr) = cb->src; 1137 *(__le16 *)(DN_SK(newsk)->addr.sdn_add.a_addr) = cb->dst; 1138 1139 menuver = *skb->data; 1140 skb_pull(skb, 1); 1141 1142 if (menuver & DN_MENUVER_ACC) 1143 dn_access_copy(skb, &(DN_SK(newsk)->accessdata)); 1144 1145 if (menuver & DN_MENUVER_USR) 1146 dn_user_copy(skb, &(DN_SK(newsk)->conndata_in)); 1147 1148 if (menuver & DN_MENUVER_PRX) 1149 DN_SK(newsk)->peer.sdn_flags |= SDF_PROXY; 1150 1151 if (menuver & DN_MENUVER_UIC) 1152 DN_SK(newsk)->peer.sdn_flags |= SDF_UICPROXY; 1153 1154 kfree_skb(skb); 1155 1156 memcpy(&(DN_SK(newsk)->conndata_out), &(DN_SK(sk)->conndata_out), 1157 sizeof(struct optdata_dn)); 1158 memcpy(&(DN_SK(newsk)->discdata_out), &(DN_SK(sk)->discdata_out), 1159 sizeof(struct optdata_dn)); 1160 1161 lock_sock(newsk); 1162 err = dn_hash_sock(newsk); 1163 if (err == 0) { 1164 sock_reset_flag(newsk, SOCK_ZAPPED); 1165 dn_send_conn_ack(newsk); 1166 1167 /* 1168 * Here we use sk->sk_allocation since although the conn conf is 1169 * for the newsk, the context is the old socket. 1170 */ 1171 if (DN_SK(newsk)->accept_mode == ACC_IMMED) 1172 err = dn_confirm_accept(newsk, &timeo, 1173 sk->sk_allocation); 1174 } 1175 release_sock(newsk); 1176 return err; 1177} 1178 1179 1180static int dn_getname(struct socket *sock, struct sockaddr *uaddr,int *uaddr_len,int peer) 1181{ 1182 struct sockaddr_dn *sa = (struct sockaddr_dn *)uaddr; 1183 struct sock *sk = sock->sk; 1184 struct dn_scp *scp = DN_SK(sk); 1185 1186 *uaddr_len = sizeof(struct sockaddr_dn); 1187 1188 lock_sock(sk); 1189 1190 if (peer) { 1191 if ((sock->state != SS_CONNECTED && 1192 sock->state != SS_CONNECTING) && 1193 scp->accept_mode == ACC_IMMED) { 1194 release_sock(sk); 1195 return -ENOTCONN; 1196 } 1197 1198 memcpy(sa, &scp->peer, sizeof(struct sockaddr_dn)); 1199 } else { 1200 memcpy(sa, &scp->addr, sizeof(struct sockaddr_dn)); 1201 } 1202 1203 release_sock(sk); 1204 1205 return 0; 1206} 1207 1208 1209static unsigned int dn_poll(struct file *file, struct socket *sock, poll_table *wait) 1210{ 1211 struct sock *sk = sock->sk; 1212 struct dn_scp *scp = DN_SK(sk); 1213 int mask = datagram_poll(file, sock, wait); 1214 1215 if (!skb_queue_empty(&scp->other_receive_queue)) 1216 mask |= POLLRDBAND; 1217 1218 return mask; 1219} 1220 1221static int dn_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg) 1222{ 1223 struct sock *sk = sock->sk; 1224 struct dn_scp *scp = DN_SK(sk); 1225 int err = -EOPNOTSUPP; 1226 long amount = 0; 1227 struct sk_buff *skb; 1228 int val; 1229 1230 switch(cmd) 1231 { 1232 case SIOCGIFADDR: 1233 case SIOCSIFADDR: 1234 return dn_dev_ioctl(cmd, (void __user *)arg); 1235 1236 case SIOCATMARK: 1237 lock_sock(sk); 1238 val = !skb_queue_empty(&scp->other_receive_queue); 1239 if (scp->state != DN_RUN) 1240 val = -ENOTCONN; 1241 release_sock(sk); 1242 return val; 1243 1244 case TIOCOUTQ: 1245 amount = sk->sk_sndbuf - sk_wmem_alloc_get(sk); 1246 if (amount < 0) 1247 amount = 0; 1248 err = put_user(amount, (int __user *)arg); 1249 break; 1250 1251 case TIOCINQ: 1252 lock_sock(sk); 1253 skb = skb_peek(&scp->other_receive_queue); 1254 if (skb) { 1255 amount = skb->len; 1256 } else { 1257 skb_queue_walk(&sk->sk_receive_queue, skb) 1258 amount += skb->len; 1259 } 1260 release_sock(sk); 1261 err = put_user(amount, (int __user *)arg); 1262 break; 1263 1264 default: 1265 err = -ENOIOCTLCMD; 1266 break; 1267 } 1268 1269 return err; 1270} 1271 1272static int dn_listen(struct socket *sock, int backlog) 1273{ 1274 struct sock *sk = sock->sk; 1275 int err = -EINVAL; 1276 1277 lock_sock(sk); 1278 1279 if (sock_flag(sk, SOCK_ZAPPED)) 1280 goto out; 1281 1282 if ((DN_SK(sk)->state != DN_O) || (sk->sk_state == TCP_LISTEN)) 1283 goto out; 1284 1285 sk->sk_max_ack_backlog = backlog; 1286 sk->sk_ack_backlog = 0; 1287 sk->sk_state = TCP_LISTEN; 1288 err = 0; 1289 dn_rehash_sock(sk); 1290 1291out: 1292 release_sock(sk); 1293 1294 return err; 1295} 1296 1297 1298static int dn_shutdown(struct socket *sock, int how) 1299{ 1300 struct sock *sk = sock->sk; 1301 struct dn_scp *scp = DN_SK(sk); 1302 int err = -ENOTCONN; 1303 1304 lock_sock(sk); 1305 1306 if (sock->state == SS_UNCONNECTED) 1307 goto out; 1308 1309 err = 0; 1310 if (sock->state == SS_DISCONNECTING) 1311 goto out; 1312 1313 err = -EINVAL; 1314 if (scp->state == DN_O) 1315 goto out; 1316 1317 if (how != SHUTDOWN_MASK) 1318 goto out; 1319 1320 sk->sk_shutdown = how; 1321 dn_destroy_sock(sk); 1322 err = 0; 1323 1324out: 1325 release_sock(sk); 1326 1327 return err; 1328} 1329 1330static int dn_setsockopt(struct socket *sock, int level, int optname, char __user *optval, unsigned int optlen) 1331{ 1332 struct sock *sk = sock->sk; 1333 int err; 1334 1335 lock_sock(sk); 1336 err = __dn_setsockopt(sock, level, optname, optval, optlen, 0); 1337 release_sock(sk); 1338 1339 return err; 1340} 1341 1342static int __dn_setsockopt(struct socket *sock, int level,int optname, char __user *optval, unsigned int optlen, int flags) 1343{ 1344 struct sock *sk = sock->sk; 1345 struct dn_scp *scp = DN_SK(sk); 1346 long timeo; 1347 union { 1348 struct optdata_dn opt; 1349 struct accessdata_dn acc; 1350 int mode; 1351 unsigned long win; 1352 int val; 1353 unsigned char services; 1354 unsigned char info; 1355 } u; 1356 int err; 1357 1358 if (optlen && !optval) 1359 return -EINVAL; 1360 1361 if (optlen > sizeof(u)) 1362 return -EINVAL; 1363 1364 if (copy_from_user(&u, optval, optlen)) 1365 return -EFAULT; 1366 1367 switch (optname) { 1368 case DSO_CONDATA: 1369 if (sock->state == SS_CONNECTED) 1370 return -EISCONN; 1371 if ((scp->state != DN_O) && (scp->state != DN_CR)) 1372 return -EINVAL; 1373 1374 if (optlen != sizeof(struct optdata_dn)) 1375 return -EINVAL; 1376 1377 if (le16_to_cpu(u.opt.opt_optl) > 16) 1378 return -EINVAL; 1379 1380 memcpy(&scp->conndata_out, &u.opt, optlen); 1381 break; 1382 1383 case DSO_DISDATA: 1384 if (sock->state != SS_CONNECTED && 1385 scp->accept_mode == ACC_IMMED) 1386 return -ENOTCONN; 1387 1388 if (optlen != sizeof(struct optdata_dn)) 1389 return -EINVAL; 1390 1391 if (le16_to_cpu(u.opt.opt_optl) > 16) 1392 return -EINVAL; 1393 1394 memcpy(&scp->discdata_out, &u.opt, optlen); 1395 break; 1396 1397 case DSO_CONACCESS: 1398 if (sock->state == SS_CONNECTED) 1399 return -EISCONN; 1400 if (scp->state != DN_O) 1401 return -EINVAL; 1402 1403 if (optlen != sizeof(struct accessdata_dn)) 1404 return -EINVAL; 1405 1406 if ((u.acc.acc_accl > DN_MAXACCL) || 1407 (u.acc.acc_passl > DN_MAXACCL) || 1408 (u.acc.acc_userl > DN_MAXACCL)) 1409 return -EINVAL; 1410 1411 memcpy(&scp->accessdata, &u.acc, optlen); 1412 break; 1413 1414 case DSO_ACCEPTMODE: 1415 if (sock->state == SS_CONNECTED) 1416 return -EISCONN; 1417 if (scp->state != DN_O) 1418 return -EINVAL; 1419 1420 if (optlen != sizeof(int)) 1421 return -EINVAL; 1422 1423 if ((u.mode != ACC_IMMED) && (u.mode != ACC_DEFER)) 1424 return -EINVAL; 1425 1426 scp->accept_mode = (unsigned char)u.mode; 1427 break; 1428 1429 case DSO_CONACCEPT: 1430 if (scp->state != DN_CR) 1431 return -EINVAL; 1432 timeo = sock_rcvtimeo(sk, 0); 1433 err = dn_confirm_accept(sk, &timeo, sk->sk_allocation); 1434 return err; 1435 1436 case DSO_CONREJECT: 1437 if (scp->state != DN_CR) 1438 return -EINVAL; 1439 1440 scp->state = DN_DR; 1441 sk->sk_shutdown = SHUTDOWN_MASK; 1442 dn_nsp_send_disc(sk, 0x38, 0, sk->sk_allocation); 1443 break; 1444 1445 default: 1446#ifdef CONFIG_NETFILTER 1447 return nf_setsockopt(sk, PF_DECnet, optname, optval, optlen); 1448#endif 1449 case DSO_LINKINFO: 1450 case DSO_STREAM: 1451 case DSO_SEQPACKET: 1452 return -ENOPROTOOPT; 1453 1454 case DSO_MAXWINDOW: 1455 if (optlen != sizeof(unsigned long)) 1456 return -EINVAL; 1457 if (u.win > NSP_MAX_WINDOW) 1458 u.win = NSP_MAX_WINDOW; 1459 if (u.win == 0) 1460 return -EINVAL; 1461 scp->max_window = u.win; 1462 if (scp->snd_window > u.win) 1463 scp->snd_window = u.win; 1464 break; 1465 1466 case DSO_NODELAY: 1467 if (optlen != sizeof(int)) 1468 return -EINVAL; 1469 if (scp->nonagle == 2) 1470 return -EINVAL; 1471 scp->nonagle = (u.val == 0) ? 0 : 1; 1472 /* if (scp->nonagle == 1) { Push pending frames } */ 1473 break; 1474 1475 case DSO_CORK: 1476 if (optlen != sizeof(int)) 1477 return -EINVAL; 1478 if (scp->nonagle == 1) 1479 return -EINVAL; 1480 scp->nonagle = (u.val == 0) ? 0 : 2; 1481 /* if (scp->nonagle == 0) { Push pending frames } */ 1482 break; 1483 1484 case DSO_SERVICES: 1485 if (optlen != sizeof(unsigned char)) 1486 return -EINVAL; 1487 if ((u.services & ~NSP_FC_MASK) != 0x01) 1488 return -EINVAL; 1489 if ((u.services & NSP_FC_MASK) == NSP_FC_MASK) 1490 return -EINVAL; 1491 scp->services_loc = u.services; 1492 break; 1493 1494 case DSO_INFO: 1495 if (optlen != sizeof(unsigned char)) 1496 return -EINVAL; 1497 if (u.info & 0xfc) 1498 return -EINVAL; 1499 scp->info_loc = u.info; 1500 break; 1501 } 1502 1503 return 0; 1504} 1505 1506static int dn_getsockopt(struct socket *sock, int level, int optname, char __user *optval, int __user *optlen) 1507{ 1508 struct sock *sk = sock->sk; 1509 int err; 1510 1511 lock_sock(sk); 1512 err = __dn_getsockopt(sock, level, optname, optval, optlen, 0); 1513 release_sock(sk); 1514 1515 return err; 1516} 1517 1518static int __dn_getsockopt(struct socket *sock, int level,int optname, char __user *optval,int __user *optlen, int flags) 1519{ 1520 struct sock *sk = sock->sk; 1521 struct dn_scp *scp = DN_SK(sk); 1522 struct linkinfo_dn link; 1523 unsigned int r_len; 1524 void *r_data = NULL; 1525 unsigned int val; 1526 1527 if(get_user(r_len , optlen)) 1528 return -EFAULT; 1529 1530 switch (optname) { 1531 case DSO_CONDATA: 1532 if (r_len > sizeof(struct optdata_dn)) 1533 r_len = sizeof(struct optdata_dn); 1534 r_data = &scp->conndata_in; 1535 break; 1536 1537 case DSO_DISDATA: 1538 if (r_len > sizeof(struct optdata_dn)) 1539 r_len = sizeof(struct optdata_dn); 1540 r_data = &scp->discdata_in; 1541 break; 1542 1543 case DSO_CONACCESS: 1544 if (r_len > sizeof(struct accessdata_dn)) 1545 r_len = sizeof(struct accessdata_dn); 1546 r_data = &scp->accessdata; 1547 break; 1548 1549 case DSO_ACCEPTMODE: 1550 if (r_len > sizeof(unsigned char)) 1551 r_len = sizeof(unsigned char); 1552 r_data = &scp->accept_mode; 1553 break; 1554 1555 case DSO_LINKINFO: 1556 if (r_len > sizeof(struct linkinfo_dn)) 1557 r_len = sizeof(struct linkinfo_dn); 1558 1559 memset(&link, 0, sizeof(link)); 1560 1561 switch (sock->state) { 1562 case SS_CONNECTING: 1563 link.idn_linkstate = LL_CONNECTING; 1564 break; 1565 case SS_DISCONNECTING: 1566 link.idn_linkstate = LL_DISCONNECTING; 1567 break; 1568 case SS_CONNECTED: 1569 link.idn_linkstate = LL_RUNNING; 1570 break; 1571 default: 1572 link.idn_linkstate = LL_INACTIVE; 1573 } 1574 1575 link.idn_segsize = scp->segsize_rem; 1576 r_data = &link; 1577 break; 1578 1579 default: 1580#ifdef CONFIG_NETFILTER 1581 { 1582 int ret, len; 1583 1584 if (get_user(len, optlen)) 1585 return -EFAULT; 1586 1587 ret = nf_getsockopt(sk, PF_DECnet, optname, optval, &len); 1588 if (ret >= 0) 1589 ret = put_user(len, optlen); 1590 return ret; 1591 } 1592#endif 1593 case DSO_STREAM: 1594 case DSO_SEQPACKET: 1595 case DSO_CONACCEPT: 1596 case DSO_CONREJECT: 1597 return -ENOPROTOOPT; 1598 1599 case DSO_MAXWINDOW: 1600 if (r_len > sizeof(unsigned long)) 1601 r_len = sizeof(unsigned long); 1602 r_data = &scp->max_window; 1603 break; 1604 1605 case DSO_NODELAY: 1606 if (r_len > sizeof(int)) 1607 r_len = sizeof(int); 1608 val = (scp->nonagle == 1); 1609 r_data = &val; 1610 break; 1611 1612 case DSO_CORK: 1613 if (r_len > sizeof(int)) 1614 r_len = sizeof(int); 1615 val = (scp->nonagle == 2); 1616 r_data = &val; 1617 break; 1618 1619 case DSO_SERVICES: 1620 if (r_len > sizeof(unsigned char)) 1621 r_len = sizeof(unsigned char); 1622 r_data = &scp->services_rem; 1623 break; 1624 1625 case DSO_INFO: 1626 if (r_len > sizeof(unsigned char)) 1627 r_len = sizeof(unsigned char); 1628 r_data = &scp->info_rem; 1629 break; 1630 } 1631 1632 if (r_data) { 1633 if (copy_to_user(optval, r_data, r_len)) 1634 return -EFAULT; 1635 if (put_user(r_len, optlen)) 1636 return -EFAULT; 1637 } 1638 1639 return 0; 1640} 1641 1642 1643static int dn_data_ready(struct sock *sk, struct sk_buff_head *q, int flags, int target) 1644{ 1645 struct sk_buff *skb; 1646 int len = 0; 1647 1648 if (flags & MSG_OOB) 1649 return !skb_queue_empty(q) ? 1 : 0; 1650 1651 skb_queue_walk(q, skb) { 1652 struct dn_skb_cb *cb = DN_SKB_CB(skb); 1653 len += skb->len; 1654 1655 if (cb->nsp_flags & 0x40) { 1656 /* SOCK_SEQPACKET reads to EOM */ 1657 if (sk->sk_type == SOCK_SEQPACKET) 1658 return 1; 1659 /* so does SOCK_STREAM unless WAITALL is specified */ 1660 if (!(flags & MSG_WAITALL)) 1661 return 1; 1662 } 1663 1664 /* minimum data length for read exceeded */ 1665 if (len >= target) 1666 return 1; 1667 } 1668 1669 return 0; 1670} 1671 1672 1673static int dn_recvmsg(struct kiocb *iocb, struct socket *sock, 1674 struct msghdr *msg, size_t size, int flags) 1675{ 1676 struct sock *sk = sock->sk; 1677 struct dn_scp *scp = DN_SK(sk); 1678 struct sk_buff_head *queue = &sk->sk_receive_queue; 1679 size_t target = size > 1 ? 1 : 0; 1680 size_t copied = 0; 1681 int rv = 0; 1682 struct sk_buff *skb, *n; 1683 struct dn_skb_cb *cb = NULL; 1684 unsigned char eor = 0; 1685 long timeo = sock_rcvtimeo(sk, flags & MSG_DONTWAIT); 1686 1687 lock_sock(sk); 1688 1689 if (sock_flag(sk, SOCK_ZAPPED)) { 1690 rv = -EADDRNOTAVAIL; 1691 goto out; 1692 } 1693 1694 if (sk->sk_shutdown & RCV_SHUTDOWN) { 1695 rv = 0; 1696 goto out; 1697 } 1698 1699 rv = dn_check_state(sk, NULL, 0, &timeo, flags); 1700 if (rv) 1701 goto out; 1702 1703 if (flags & ~(MSG_CMSG_COMPAT|MSG_PEEK|MSG_OOB|MSG_WAITALL|MSG_DONTWAIT|MSG_NOSIGNAL)) { 1704 rv = -EOPNOTSUPP; 1705 goto out; 1706 } 1707 1708 if (flags & MSG_OOB) 1709 queue = &scp->other_receive_queue; 1710 1711 if (flags & MSG_WAITALL) 1712 target = size; 1713 1714 1715 /* 1716 * See if there is data ready to read, sleep if there isn't 1717 */ 1718 for(;;) { 1719 DEFINE_WAIT(wait); 1720 1721 if (sk->sk_err) 1722 goto out; 1723 1724 if (!skb_queue_empty(&scp->other_receive_queue)) { 1725 if (!(flags & MSG_OOB)) { 1726 msg->msg_flags |= MSG_OOB; 1727 if (!scp->other_report) { 1728 scp->other_report = 1; 1729 goto out; 1730 } 1731 } 1732 } 1733 1734 if (scp->state != DN_RUN) 1735 goto out; 1736 1737 if (signal_pending(current)) { 1738 rv = sock_intr_errno(timeo); 1739 goto out; 1740 } 1741 1742 if (dn_data_ready(sk, queue, flags, target)) 1743 break; 1744 1745 if (flags & MSG_DONTWAIT) { 1746 rv = -EWOULDBLOCK; 1747 goto out; 1748 } 1749 1750 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 1751 set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); 1752 sk_wait_event(sk, &timeo, dn_data_ready(sk, queue, flags, target)); 1753 clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); 1754 finish_wait(sk_sleep(sk), &wait); 1755 } 1756 1757 skb_queue_walk_safe(queue, skb, n) { 1758 unsigned int chunk = skb->len; 1759 cb = DN_SKB_CB(skb); 1760 1761 if ((chunk + copied) > size) 1762 chunk = size - copied; 1763 1764 if (memcpy_toiovec(msg->msg_iov, skb->data, chunk)) { 1765 rv = -EFAULT; 1766 break; 1767 } 1768 copied += chunk; 1769 1770 if (!(flags & MSG_PEEK)) 1771 skb_pull(skb, chunk); 1772 1773 eor = cb->nsp_flags & 0x40; 1774 1775 if (skb->len == 0) { 1776 skb_unlink(skb, queue); 1777 kfree_skb(skb); 1778 /* 1779 * N.B. Don't refer to skb or cb after this point 1780 * in loop. 1781 */ 1782 if ((scp->flowloc_sw == DN_DONTSEND) && !dn_congested(sk)) { 1783 scp->flowloc_sw = DN_SEND; 1784 dn_nsp_send_link(sk, DN_SEND, 0); 1785 } 1786 } 1787 1788 if (eor) { 1789 if (sk->sk_type == SOCK_SEQPACKET) 1790 break; 1791 if (!(flags & MSG_WAITALL)) 1792 break; 1793 } 1794 1795 if (flags & MSG_OOB) 1796 break; 1797 1798 if (copied >= target) 1799 break; 1800 } 1801 1802 rv = copied; 1803 1804 1805 if (eor && (sk->sk_type == SOCK_SEQPACKET)) 1806 msg->msg_flags |= MSG_EOR; 1807 1808out: 1809 if (rv == 0) 1810 rv = (flags & MSG_PEEK) ? -sk->sk_err : sock_error(sk); 1811 1812 if ((rv >= 0) && msg->msg_name) { 1813 memcpy(msg->msg_name, &scp->peer, sizeof(struct sockaddr_dn)); 1814 msg->msg_namelen = sizeof(struct sockaddr_dn); 1815 } 1816 1817 release_sock(sk); 1818 1819 return rv; 1820} 1821 1822 1823static inline int dn_queue_too_long(struct dn_scp *scp, struct sk_buff_head *queue, int flags) 1824{ 1825 unsigned char fctype = scp->services_rem & NSP_FC_MASK; 1826 if (skb_queue_len(queue) >= scp->snd_window) 1827 return 1; 1828 if (fctype != NSP_FC_NONE) { 1829 if (flags & MSG_OOB) { 1830 if (scp->flowrem_oth == 0) 1831 return 1; 1832 } else { 1833 if (scp->flowrem_dat == 0) 1834 return 1; 1835 } 1836 } 1837 return 0; 1838} 1839 1840/* 1841 * The DECnet spec requires that the "routing layer" accepts packets which 1842 * are at least 230 bytes in size. This excludes any headers which the NSP 1843 * layer might add, so we always assume that we'll be using the maximal 1844 * length header on data packets. The variation in length is due to the 1845 * inclusion (or not) of the two 16 bit acknowledgement fields so it doesn't 1846 * make much practical difference. 1847 */ 1848unsigned dn_mss_from_pmtu(struct net_device *dev, int mtu) 1849{ 1850 unsigned mss = 230 - DN_MAX_NSP_DATA_HEADER; 1851 if (dev) { 1852 struct dn_dev *dn_db = rcu_dereference_raw(dev->dn_ptr); 1853 mtu -= LL_RESERVED_SPACE(dev); 1854 if (dn_db->use_long) 1855 mtu -= 21; 1856 else 1857 mtu -= 6; 1858 mtu -= DN_MAX_NSP_DATA_HEADER; 1859 } else { 1860 /* 1861 * 21 = long header, 16 = guess at MAC header length 1862 */ 1863 mtu -= (21 + DN_MAX_NSP_DATA_HEADER + 16); 1864 } 1865 if (mtu > mss) 1866 mss = mtu; 1867 return mss; 1868} 1869 1870static inline unsigned int dn_current_mss(struct sock *sk, int flags) 1871{ 1872 struct dst_entry *dst = __sk_dst_get(sk); 1873 struct dn_scp *scp = DN_SK(sk); 1874 int mss_now = min_t(int, scp->segsize_loc, scp->segsize_rem); 1875 1876 /* Other data messages are limited to 16 bytes per packet */ 1877 if (flags & MSG_OOB) 1878 return 16; 1879 1880 /* This works out the maximum size of segment we can send out */ 1881 if (dst) { 1882 u32 mtu = dst_mtu(dst); 1883 mss_now = min_t(int, dn_mss_from_pmtu(dst->dev, mtu), mss_now); 1884 } 1885 1886 return mss_now; 1887} 1888 1889/* 1890 * N.B. We get the timeout wrong here, but then we always did get it 1891 * wrong before and this is another step along the road to correcting 1892 * it. It ought to get updated each time we pass through the routine, 1893 * but in practise it probably doesn't matter too much for now. 1894 */ 1895static inline struct sk_buff *dn_alloc_send_pskb(struct sock *sk, 1896 unsigned long datalen, int noblock, 1897 int *errcode) 1898{ 1899 struct sk_buff *skb = sock_alloc_send_skb(sk, datalen, 1900 noblock, errcode); 1901 if (skb) { 1902 skb->protocol = htons(ETH_P_DNA_RT); 1903 skb->pkt_type = PACKET_OUTGOING; 1904 } 1905 return skb; 1906} 1907 1908static int dn_sendmsg(struct kiocb *iocb, struct socket *sock, 1909 struct msghdr *msg, size_t size) 1910{ 1911 struct sock *sk = sock->sk; 1912 struct dn_scp *scp = DN_SK(sk); 1913 size_t mss; 1914 struct sk_buff_head *queue = &scp->data_xmit_queue; 1915 int flags = msg->msg_flags; 1916 int err = 0; 1917 size_t sent = 0; 1918 int addr_len = msg->msg_namelen; 1919 struct sockaddr_dn *addr = (struct sockaddr_dn *)msg->msg_name; 1920 struct sk_buff *skb = NULL; 1921 struct dn_skb_cb *cb; 1922 size_t len; 1923 unsigned char fctype; 1924 long timeo; 1925 1926 if (flags & ~(MSG_TRYHARD|MSG_OOB|MSG_DONTWAIT|MSG_EOR|MSG_NOSIGNAL|MSG_MORE|MSG_CMSG_COMPAT)) 1927 return -EOPNOTSUPP; 1928 1929 if (addr_len && (addr_len != sizeof(struct sockaddr_dn))) 1930 return -EINVAL; 1931 1932 lock_sock(sk); 1933 timeo = sock_sndtimeo(sk, flags & MSG_DONTWAIT); 1934 /* 1935 * The only difference between stream sockets and sequenced packet 1936 * sockets is that the stream sockets always behave as if MSG_EOR 1937 * has been set. 1938 */ 1939 if (sock->type == SOCK_STREAM) { 1940 if (flags & MSG_EOR) { 1941 err = -EINVAL; 1942 goto out; 1943 } 1944 flags |= MSG_EOR; 1945 } 1946 1947 1948 err = dn_check_state(sk, addr, addr_len, &timeo, flags); 1949 if (err) 1950 goto out_err; 1951 1952 if (sk->sk_shutdown & SEND_SHUTDOWN) { 1953 err = -EPIPE; 1954 if (!(flags & MSG_NOSIGNAL)) 1955 send_sig(SIGPIPE, current, 0); 1956 goto out_err; 1957 } 1958 1959 if ((flags & MSG_TRYHARD) && sk->sk_dst_cache) 1960 dst_negative_advice(sk); 1961 1962 mss = scp->segsize_rem; 1963 fctype = scp->services_rem & NSP_FC_MASK; 1964 1965 mss = dn_current_mss(sk, flags); 1966 1967 if (flags & MSG_OOB) { 1968 queue = &scp->other_xmit_queue; 1969 if (size > mss) { 1970 err = -EMSGSIZE; 1971 goto out; 1972 } 1973 } 1974 1975 scp->persist_fxn = dn_nsp_xmit_timeout; 1976 1977 while(sent < size) { 1978 err = sock_error(sk); 1979 if (err) 1980 goto out; 1981 1982 if (signal_pending(current)) { 1983 err = sock_intr_errno(timeo); 1984 goto out; 1985 } 1986 1987 /* 1988 * Calculate size that we wish to send. 1989 */ 1990 len = size - sent; 1991 1992 if (len > mss) 1993 len = mss; 1994 1995 /* 1996 * Wait for queue size to go down below the window 1997 * size. 1998 */ 1999 if (dn_queue_too_long(scp, queue, flags)) { 2000 DEFINE_WAIT(wait); 2001 2002 if (flags & MSG_DONTWAIT) { 2003 err = -EWOULDBLOCK; 2004 goto out; 2005 } 2006 2007 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); 2008 set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); 2009 sk_wait_event(sk, &timeo, 2010 !dn_queue_too_long(scp, queue, flags)); 2011 clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); 2012 finish_wait(sk_sleep(sk), &wait); 2013 continue; 2014 } 2015 2016 /* 2017 * Get a suitably sized skb. 2018 * 64 is a bit of a hack really, but its larger than any 2019 * link-layer headers and has served us well as a good 2020 * guess as to their real length. 2021 */ 2022 skb = dn_alloc_send_pskb(sk, len + 64 + DN_MAX_NSP_DATA_HEADER, 2023 flags & MSG_DONTWAIT, &err); 2024 2025 if (err) 2026 break; 2027 2028 if (!skb) 2029 continue; 2030 2031 cb = DN_SKB_CB(skb); 2032 2033 skb_reserve(skb, 64 + DN_MAX_NSP_DATA_HEADER); 2034 2035 if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)) { 2036 err = -EFAULT; 2037 goto out; 2038 } 2039 2040 if (flags & MSG_OOB) { 2041 cb->nsp_flags = 0x30; 2042 if (fctype != NSP_FC_NONE) 2043 scp->flowrem_oth--; 2044 } else { 2045 cb->nsp_flags = 0x00; 2046 if (scp->seg_total == 0) 2047 cb->nsp_flags |= 0x20; 2048 2049 scp->seg_total += len; 2050 2051 if (((sent + len) == size) && (flags & MSG_EOR)) { 2052 cb->nsp_flags |= 0x40; 2053 scp->seg_total = 0; 2054 if (fctype == NSP_FC_SCMC) 2055 scp->flowrem_dat--; 2056 } 2057 if (fctype == NSP_FC_SRC) 2058 scp->flowrem_dat--; 2059 } 2060 2061 sent += len; 2062 dn_nsp_queue_xmit(sk, skb, sk->sk_allocation, flags & MSG_OOB); 2063 skb = NULL; 2064 2065 scp->persist = dn_nsp_persist(sk); 2066 2067 } 2068out: 2069 2070 kfree_skb(skb); 2071 2072 release_sock(sk); 2073 2074 return sent ? sent : err; 2075 2076out_err: 2077 err = sk_stream_error(sk, flags, err); 2078 release_sock(sk); 2079 return err; 2080} 2081 2082static int dn_device_event(struct notifier_block *this, unsigned long event, 2083 void *ptr) 2084{ 2085 struct net_device *dev = (struct net_device *)ptr; 2086 2087 if (!net_eq(dev_net(dev), &init_net)) 2088 return NOTIFY_DONE; 2089 2090 switch (event) { 2091 case NETDEV_UP: 2092 dn_dev_up(dev); 2093 break; 2094 case NETDEV_DOWN: 2095 dn_dev_down(dev); 2096 break; 2097 default: 2098 break; 2099 } 2100 2101 return NOTIFY_DONE; 2102} 2103 2104static struct notifier_block dn_dev_notifier = { 2105 .notifier_call = dn_device_event, 2106}; 2107 2108extern int dn_route_rcv(struct sk_buff *, struct net_device *, struct packet_type *, struct net_device *); 2109 2110static struct packet_type dn_dix_packet_type __read_mostly = { 2111 .type = cpu_to_be16(ETH_P_DNA_RT), 2112 .func = dn_route_rcv, 2113}; 2114 2115#ifdef CONFIG_PROC_FS 2116struct dn_iter_state { 2117 int bucket; 2118}; 2119 2120static struct sock *dn_socket_get_first(struct seq_file *seq) 2121{ 2122 struct dn_iter_state *state = seq->private; 2123 struct sock *n = NULL; 2124 2125 for(state->bucket = 0; 2126 state->bucket < DN_SK_HASH_SIZE; 2127 ++state->bucket) { 2128 n = sk_head(&dn_sk_hash[state->bucket]); 2129 if (n) 2130 break; 2131 } 2132 2133 return n; 2134} 2135 2136static struct sock *dn_socket_get_next(struct seq_file *seq, 2137 struct sock *n) 2138{ 2139 struct dn_iter_state *state = seq->private; 2140 2141 n = sk_next(n); 2142try_again: 2143 if (n) 2144 goto out; 2145 if (++state->bucket >= DN_SK_HASH_SIZE) 2146 goto out; 2147 n = sk_head(&dn_sk_hash[state->bucket]); 2148 goto try_again; 2149out: 2150 return n; 2151} 2152 2153static struct sock *socket_get_idx(struct seq_file *seq, loff_t *pos) 2154{ 2155 struct sock *sk = dn_socket_get_first(seq); 2156 2157 if (sk) { 2158 while(*pos && (sk = dn_socket_get_next(seq, sk))) 2159 --*pos; 2160 } 2161 return *pos ? NULL : sk; 2162} 2163 2164static void *dn_socket_get_idx(struct seq_file *seq, loff_t pos) 2165{ 2166 void *rc; 2167 read_lock_bh(&dn_hash_lock); 2168 rc = socket_get_idx(seq, &pos); 2169 if (!rc) { 2170 read_unlock_bh(&dn_hash_lock); 2171 } 2172 return rc; 2173} 2174 2175static void *dn_socket_seq_start(struct seq_file *seq, loff_t *pos) 2176{ 2177 return *pos ? dn_socket_get_idx(seq, *pos - 1) : SEQ_START_TOKEN; 2178} 2179 2180static void *dn_socket_seq_next(struct seq_file *seq, void *v, loff_t *pos) 2181{ 2182 void *rc; 2183 2184 if (v == SEQ_START_TOKEN) { 2185 rc = dn_socket_get_idx(seq, 0); 2186 goto out; 2187 } 2188 2189 rc = dn_socket_get_next(seq, v); 2190 if (rc) 2191 goto out; 2192 read_unlock_bh(&dn_hash_lock); 2193out: 2194 ++*pos; 2195 return rc; 2196} 2197 2198static void dn_socket_seq_stop(struct seq_file *seq, void *v) 2199{ 2200 if (v && v != SEQ_START_TOKEN) 2201 read_unlock_bh(&dn_hash_lock); 2202} 2203 2204#define IS_NOT_PRINTABLE(x) ((x) < 32 || (x) > 126) 2205 2206static void dn_printable_object(struct sockaddr_dn *dn, unsigned char *buf) 2207{ 2208 int i; 2209 2210 switch (le16_to_cpu(dn->sdn_objnamel)) { 2211 case 0: 2212 sprintf(buf, "%d", dn->sdn_objnum); 2213 break; 2214 default: 2215 for (i = 0; i < le16_to_cpu(dn->sdn_objnamel); i++) { 2216 buf[i] = dn->sdn_objname[i]; 2217 if (IS_NOT_PRINTABLE(buf[i])) 2218 buf[i] = '.'; 2219 } 2220 buf[i] = 0; 2221 } 2222} 2223 2224static char *dn_state2asc(unsigned char state) 2225{ 2226 switch (state) { 2227 case DN_O: 2228 return "OPEN"; 2229 case DN_CR: 2230 return " CR"; 2231 case DN_DR: 2232 return " DR"; 2233 case DN_DRC: 2234 return " DRC"; 2235 case DN_CC: 2236 return " CC"; 2237 case DN_CI: 2238 return " CI"; 2239 case DN_NR: 2240 return " NR"; 2241 case DN_NC: 2242 return " NC"; 2243 case DN_CD: 2244 return " CD"; 2245 case DN_RJ: 2246 return " RJ"; 2247 case DN_RUN: 2248 return " RUN"; 2249 case DN_DI: 2250 return " DI"; 2251 case DN_DIC: 2252 return " DIC"; 2253 case DN_DN: 2254 return " DN"; 2255 case DN_CL: 2256 return " CL"; 2257 case DN_CN: 2258 return " CN"; 2259 } 2260 2261 return "????"; 2262} 2263 2264static inline void dn_socket_format_entry(struct seq_file *seq, struct sock *sk) 2265{ 2266 struct dn_scp *scp = DN_SK(sk); 2267 char buf1[DN_ASCBUF_LEN]; 2268 char buf2[DN_ASCBUF_LEN]; 2269 char local_object[DN_MAXOBJL+3]; 2270 char remote_object[DN_MAXOBJL+3]; 2271 2272 dn_printable_object(&scp->addr, local_object); 2273 dn_printable_object(&scp->peer, remote_object); 2274 2275 seq_printf(seq, 2276 "%6s/%04X %04d:%04d %04d:%04d %01d %-16s " 2277 "%6s/%04X %04d:%04d %04d:%04d %01d %-16s %4s %s\n", 2278 dn_addr2asc(le16_to_cpu(dn_saddr2dn(&scp->addr)), buf1), 2279 scp->addrloc, 2280 scp->numdat, 2281 scp->numoth, 2282 scp->ackxmt_dat, 2283 scp->ackxmt_oth, 2284 scp->flowloc_sw, 2285 local_object, 2286 dn_addr2asc(le16_to_cpu(dn_saddr2dn(&scp->peer)), buf2), 2287 scp->addrrem, 2288 scp->numdat_rcv, 2289 scp->numoth_rcv, 2290 scp->ackrcv_dat, 2291 scp->ackrcv_oth, 2292 scp->flowrem_sw, 2293 remote_object, 2294 dn_state2asc(scp->state), 2295 ((scp->accept_mode == ACC_IMMED) ? "IMMED" : "DEFER")); 2296} 2297 2298static int dn_socket_seq_show(struct seq_file *seq, void *v) 2299{ 2300 if (v == SEQ_START_TOKEN) { 2301 seq_puts(seq, "Local Remote\n"); 2302 } else { 2303 dn_socket_format_entry(seq, v); 2304 } 2305 return 0; 2306} 2307 2308static const struct seq_operations dn_socket_seq_ops = { 2309 .start = dn_socket_seq_start, 2310 .next = dn_socket_seq_next, 2311 .stop = dn_socket_seq_stop, 2312 .show = dn_socket_seq_show, 2313}; 2314 2315static int dn_socket_seq_open(struct inode *inode, struct file *file) 2316{ 2317 return seq_open_private(file, &dn_socket_seq_ops, 2318 sizeof(struct dn_iter_state)); 2319} 2320 2321static const struct file_operations dn_socket_seq_fops = { 2322 .owner = THIS_MODULE, 2323 .open = dn_socket_seq_open, 2324 .read = seq_read, 2325 .llseek = seq_lseek, 2326 .release = seq_release_private, 2327}; 2328#endif 2329 2330static const struct net_proto_family dn_family_ops = { 2331 .family = AF_DECnet, 2332 .create = dn_create, 2333 .owner = THIS_MODULE, 2334}; 2335 2336static const struct proto_ops dn_proto_ops = { 2337 .family = AF_DECnet, 2338 .owner = THIS_MODULE, 2339 .release = dn_release, 2340 .bind = dn_bind, 2341 .connect = dn_connect, 2342 .socketpair = sock_no_socketpair, 2343 .accept = dn_accept, 2344 .getname = dn_getname, 2345 .poll = dn_poll, 2346 .ioctl = dn_ioctl, 2347 .listen = dn_listen, 2348 .shutdown = dn_shutdown, 2349 .setsockopt = dn_setsockopt, 2350 .getsockopt = dn_getsockopt, 2351 .sendmsg = dn_sendmsg, 2352 .recvmsg = dn_recvmsg, 2353 .mmap = sock_no_mmap, 2354 .sendpage = sock_no_sendpage, 2355}; 2356 2357void dn_register_sysctl(void); 2358void dn_unregister_sysctl(void); 2359 2360MODULE_DESCRIPTION("The Linux DECnet Network Protocol"); 2361MODULE_AUTHOR("Linux DECnet Project Team"); 2362MODULE_LICENSE("GPL"); 2363MODULE_ALIAS_NETPROTO(PF_DECnet); 2364 2365static char banner[] __initdata = KERN_INFO "NET4: DECnet for Linux: V.2.5.68s (C) 1995-2003 Linux DECnet Project Team\n"; 2366 2367static int __init decnet_init(void) 2368{ 2369 int rc; 2370 2371 printk(banner); 2372 2373 rc = proto_register(&dn_proto, 1); 2374 if (rc != 0) 2375 goto out; 2376 2377 dn_neigh_init(); 2378 dn_dev_init(); 2379 dn_route_init(); 2380 dn_fib_init(); 2381 2382 sock_register(&dn_family_ops); 2383 dev_add_pack(&dn_dix_packet_type); 2384 register_netdevice_notifier(&dn_dev_notifier); 2385 2386 proc_net_fops_create(&init_net, "decnet", S_IRUGO, &dn_socket_seq_fops); 2387 dn_register_sysctl(); 2388out: 2389 return rc; 2390 2391} 2392module_init(decnet_init); 2393 2394/* 2395 * Prevent DECnet module unloading until its fixed properly. 2396 * Requires an audit of the code to check for memory leaks and 2397 * initialisation problems etc. 2398 */ 2399#if 0 2400static void __exit decnet_exit(void) 2401{ 2402 sock_unregister(AF_DECnet); 2403 rtnl_unregister_all(PF_DECnet); 2404 dev_remove_pack(&dn_dix_packet_type); 2405 2406 dn_unregister_sysctl(); 2407 2408 unregister_netdevice_notifier(&dn_dev_notifier); 2409 2410 dn_route_cleanup(); 2411 dn_dev_cleanup(); 2412 dn_neigh_cleanup(); 2413 dn_fib_cleanup(); 2414 2415 proc_net_remove(&init_net, "decnet"); 2416 2417 proto_unregister(&dn_proto); 2418 2419 rcu_barrier_bh(); /* Wait for completion of call_rcu_bh()'s */ 2420} 2421module_exit(decnet_exit); 2422#endif