Merge master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6

+1010 -735
+19 -13
drivers/connector/cn_queue.c
··· 31 31 #include <linux/connector.h> 32 32 #include <linux/delay.h> 33 33 34 - static void cn_queue_wrapper(void *data) 34 + void cn_queue_wrapper(void *data) 35 35 { 36 - struct cn_callback_entry *cbq = data; 36 + struct cn_callback_data *d = data; 37 37 38 - cbq->cb->callback(cbq->cb->priv); 39 - cbq->destruct_data(cbq->ddata); 40 - cbq->ddata = NULL; 38 + d->callback(d->callback_priv); 39 + 40 + d->destruct_data(d->ddata); 41 + d->ddata = NULL; 42 + 43 + kfree(d->free); 41 44 } 42 45 43 - static struct cn_callback_entry *cn_queue_alloc_callback_entry(struct cn_callback *cb) 46 + static struct cn_callback_entry *cn_queue_alloc_callback_entry(char *name, struct cb_id *id, void (*callback)(void *)) 44 47 { 45 48 struct cn_callback_entry *cbq; 46 49 ··· 53 50 return NULL; 54 51 } 55 52 56 - cbq->cb = cb; 57 - INIT_WORK(&cbq->work, &cn_queue_wrapper, cbq); 53 + snprintf(cbq->id.name, sizeof(cbq->id.name), "%s", name); 54 + memcpy(&cbq->id.id, id, sizeof(struct cb_id)); 55 + cbq->data.callback = callback; 56 + 57 + INIT_WORK(&cbq->work, &cn_queue_wrapper, &cbq->data); 58 58 return cbq; 59 59 } 60 60 ··· 74 68 return ((i1->idx == i2->idx) && (i1->val == i2->val)); 75 69 } 76 70 77 - int cn_queue_add_callback(struct cn_queue_dev *dev, struct cn_callback *cb) 71 + int cn_queue_add_callback(struct cn_queue_dev *dev, char *name, struct cb_id *id, void (*callback)(void *)) 78 72 { 79 73 struct cn_callback_entry *cbq, *__cbq; 80 74 int found = 0; 81 75 82 - cbq = cn_queue_alloc_callback_entry(cb); 76 + cbq = cn_queue_alloc_callback_entry(name, id, callback); 83 77 if (!cbq) 84 78 return -ENOMEM; 85 79 ··· 88 82 89 83 spin_lock_bh(&dev->queue_lock); 90 84 list_for_each_entry(__cbq, &dev->queue_list, callback_entry) { 91 - if (cn_cb_equal(&__cbq->cb->id, &cb->id)) { 85 + if (cn_cb_equal(&__cbq->id.id, id)) { 92 86 found = 1; 93 87 break; 94 88 } ··· 105 99 106 100 cbq->nls = dev->nls; 107 101 cbq->seq = 0; 108 - cbq->group = cbq->cb->id.idx; 102 + cbq->group = cbq->id.id.idx; 109 103 110 104 return 0; 111 105 } ··· 117 111 118 112 spin_lock_bh(&dev->queue_lock); 119 113 list_for_each_entry_safe(cbq, n, &dev->queue_list, callback_entry) { 120 - if (cn_cb_equal(&cbq->cb->id, id)) { 114 + if (cn_cb_equal(&cbq->id.id, id)) { 121 115 list_del(&cbq->callback_entry); 122 116 found = 1; 123 117 break;
+38 -36
drivers/connector/connector.c
··· 84 84 spin_lock_bh(&dev->cbdev->queue_lock); 85 85 list_for_each_entry(__cbq, &dev->cbdev->queue_list, 86 86 callback_entry) { 87 - if (cn_cb_equal(&__cbq->cb->id, &msg->id)) { 87 + if (cn_cb_equal(&__cbq->id.id, &msg->id)) { 88 88 found = 1; 89 89 group = __cbq->group; 90 90 } ··· 127 127 { 128 128 struct cn_callback_entry *__cbq; 129 129 struct cn_dev *dev = &cdev; 130 - int found = 0; 130 + int err = -ENODEV; 131 131 132 132 spin_lock_bh(&dev->cbdev->queue_lock); 133 133 list_for_each_entry(__cbq, &dev->cbdev->queue_list, callback_entry) { 134 - if (cn_cb_equal(&__cbq->cb->id, &msg->id)) { 135 - /* 136 - * Let's scream if there is some magic and the 137 - * data will arrive asynchronously here. 138 - * [i.e. netlink messages will be queued]. 139 - * After the first warning I will fix it 140 - * quickly, but now I think it is 141 - * impossible. --zbr (2004_04_27). 142 - */ 134 + if (cn_cb_equal(&__cbq->id.id, &msg->id)) { 143 135 if (likely(!test_bit(0, &__cbq->work.pending) && 144 - __cbq->ddata == NULL)) { 145 - __cbq->cb->priv = msg; 136 + __cbq->data.ddata == NULL)) { 137 + __cbq->data.callback_priv = msg; 146 138 147 - __cbq->ddata = data; 148 - __cbq->destruct_data = destruct_data; 139 + __cbq->data.ddata = data; 140 + __cbq->data.destruct_data = destruct_data; 149 141 150 142 if (queue_work(dev->cbdev->cn_queue, 151 143 &__cbq->work)) 152 - found = 1; 144 + err = 0; 153 145 } else { 154 - printk("%s: cbq->data=%p, " 155 - "work->pending=%08lx.\n", 156 - __func__, __cbq->ddata, 157 - __cbq->work.pending); 158 - WARN_ON(1); 146 + struct work_struct *w; 147 + struct cn_callback_data *d; 148 + 149 + w = kzalloc(sizeof(*w) + sizeof(*d), GFP_ATOMIC); 150 + if (w) { 151 + d = (struct cn_callback_data *)(w+1); 152 + 153 + d->callback_priv = msg; 154 + d->callback = __cbq->data.callback; 155 + d->ddata = data; 156 + d->destruct_data = destruct_data; 157 + d->free = w; 158 + 159 + INIT_LIST_HEAD(&w->entry); 160 + w->pending = 0; 161 + w->func = &cn_queue_wrapper; 162 + w->data = d; 163 + init_timer(&w->timer); 164 + 165 + if (queue_work(dev->cbdev->cn_queue, w)) 166 + err = 0; 167 + else { 168 + kfree(w); 169 + err = -EINVAL; 170 + } 171 + } else 172 + err = -ENOMEM; 159 173 } 160 174 break; 161 175 } 162 176 } 163 177 spin_unlock_bh(&dev->cbdev->queue_lock); 164 178 165 - return found ? 0 : -ENODEV; 179 + return err; 166 180 } 167 181 168 182 /* ··· 305 291 { 306 292 int err; 307 293 struct cn_dev *dev = &cdev; 308 - struct cn_callback *cb; 309 294 310 - cb = kzalloc(sizeof(*cb), GFP_KERNEL); 311 - if (!cb) 312 - return -ENOMEM; 313 - 314 - scnprintf(cb->name, sizeof(cb->name), "%s", name); 315 - 316 - memcpy(&cb->id, id, sizeof(cb->id)); 317 - cb->callback = callback; 318 - 319 - err = cn_queue_add_callback(dev->cbdev, cb); 320 - if (err) { 321 - kfree(cb); 295 + err = cn_queue_add_callback(dev->cbdev, name, id, callback); 296 + if (err) 322 297 return err; 323 - } 324 298 325 299 cn_notify(id, 0); 326 300
+15 -6
include/linux/connector.h
··· 104 104 struct sock *nls; 105 105 }; 106 106 107 - struct cn_callback { 107 + struct cn_callback_id { 108 108 unsigned char name[CN_CBQ_NAMELEN]; 109 - 110 109 struct cb_id id; 110 + }; 111 + 112 + struct cn_callback_data { 113 + void (*destruct_data) (void *); 114 + void *ddata; 115 + 116 + void *callback_priv; 111 117 void (*callback) (void *); 112 - void *priv; 118 + 119 + void *free; 113 120 }; 114 121 115 122 struct cn_callback_entry { ··· 125 118 struct work_struct work; 126 119 struct cn_queue_dev *pdev; 127 120 128 - void (*destruct_data) (void *); 129 - void *ddata; 121 + struct cn_callback_id id; 122 + struct cn_callback_data data; 130 123 131 124 int seq, group; 132 125 struct sock *nls; ··· 151 144 void cn_del_callback(struct cb_id *); 152 145 int cn_netlink_send(struct cn_msg *, u32, int); 153 146 154 - int cn_queue_add_callback(struct cn_queue_dev *dev, struct cn_callback *cb); 147 + int cn_queue_add_callback(struct cn_queue_dev *dev, char *name, struct cb_id *id, void (*callback)(void *)); 155 148 void cn_queue_del_callback(struct cn_queue_dev *dev, struct cb_id *id); 156 149 157 150 struct cn_queue_dev *cn_queue_alloc_dev(char *name, struct sock *); 158 151 void cn_queue_free_dev(struct cn_queue_dev *dev); 159 152 160 153 int cn_cb_equal(struct cb_id *, struct cb_id *); 154 + 155 + void cn_queue_wrapper(void *data); 161 156 162 157 extern int cn_already_initialized; 163 158
+5 -7
include/linux/netfilter_ipv4/ip_nat_core.h
··· 5 5 6 6 /* This header used to share core functionality between the standalone 7 7 NAT module, and the compatibility layer's use of NAT for masquerading. */ 8 - extern int ip_nat_init(void); 9 - extern void ip_nat_cleanup(void); 10 8 11 - extern unsigned int nat_packet(struct ip_conntrack *ct, 9 + extern unsigned int ip_nat_packet(struct ip_conntrack *ct, 12 10 enum ip_conntrack_info conntrackinfo, 13 11 unsigned int hooknum, 14 12 struct sk_buff **pskb); 15 13 16 - extern int icmp_reply_translation(struct sk_buff **pskb, 17 - struct ip_conntrack *ct, 18 - enum ip_nat_manip_type manip, 19 - enum ip_conntrack_dir dir); 14 + extern int ip_nat_icmp_reply_translation(struct sk_buff **pskb, 15 + struct ip_conntrack *ct, 16 + enum ip_nat_manip_type manip, 17 + enum ip_conntrack_dir dir); 20 18 #endif /* _IP_NAT_CORE_H */
+25 -1
include/linux/sysctl.h
··· 202 202 NET_TR=14, 203 203 NET_DECNET=15, 204 204 NET_ECONET=16, 205 - NET_SCTP=17, 205 + NET_SCTP=17, 206 + NET_LLC=18, 206 207 }; 207 208 208 209 /* /proc/sys/kernel/random */ ··· 523 522 NET_IPX_FORWARDING=2 524 523 }; 525 524 525 + /* /proc/sys/net/llc */ 526 + enum { 527 + NET_LLC2=1, 528 + NET_LLC_STATION=2, 529 + }; 530 + 531 + /* /proc/sys/net/llc/llc2 */ 532 + enum { 533 + NET_LLC2_TIMEOUT=1, 534 + }; 535 + 536 + /* /proc/sys/net/llc/station */ 537 + enum { 538 + NET_LLC_STATION_ACK_TIMEOUT=1, 539 + }; 540 + 541 + /* /proc/sys/net/llc/llc2/timeout */ 542 + enum { 543 + NET_LLC2_ACK_TIMEOUT=1, 544 + NET_LLC2_P_TIMEOUT=2, 545 + NET_LLC2_REJ_TIMEOUT=3, 546 + NET_LLC2_BUSY_TIMEOUT=4, 547 + }; 526 548 527 549 /* /proc/sys/net/appletalk */ 528 550 enum {
+30
include/net/llc.h
··· 17 17 #include <linux/list.h> 18 18 #include <linux/spinlock.h> 19 19 20 + #include <asm/atomic.h> 21 + 20 22 struct net_device; 21 23 struct packet_type; 22 24 struct sk_buff; ··· 46 44 unsigned char state; 47 45 unsigned char p_bit; 48 46 unsigned char f_bit; 47 + atomic_t refcnt; 49 48 int (*rcv_func)(struct sk_buff *skb, 50 49 struct net_device *dev, 51 50 struct packet_type *pt, ··· 84 81 struct net_device *dev, 85 82 struct packet_type *pt, 86 83 struct net_device *orig_dev)); 84 + static inline void llc_sap_hold(struct llc_sap *sap) 85 + { 86 + atomic_inc(&sap->refcnt); 87 + } 88 + 87 89 extern void llc_sap_close(struct llc_sap *sap); 90 + 91 + static inline void llc_sap_put(struct llc_sap *sap) 92 + { 93 + if (atomic_dec_and_test(&sap->refcnt)) 94 + llc_sap_close(sap); 95 + } 88 96 89 97 extern struct llc_sap *llc_sap_find(unsigned char sap_value); 90 98 91 99 extern int llc_build_and_send_ui_pkt(struct llc_sap *sap, struct sk_buff *skb, 92 100 unsigned char *dmac, unsigned char dsap); 101 + 102 + extern void llc_sap_handler(struct llc_sap *sap, struct sk_buff *skb); 103 + extern void llc_conn_handler(struct llc_sap *sap, struct sk_buff *skb); 93 104 94 105 extern int llc_station_init(void); 95 106 extern void llc_station_exit(void); ··· 115 98 #define llc_proc_init() (0) 116 99 #define llc_proc_exit() do { } while(0) 117 100 #endif /* CONFIG_PROC_FS */ 101 + #ifdef CONFIG_SYSCTL 102 + extern int llc_sysctl_init(void); 103 + extern void llc_sysctl_exit(void); 104 + 105 + extern int sysctl_llc2_ack_timeout; 106 + extern int sysctl_llc2_busy_timeout; 107 + extern int sysctl_llc2_p_timeout; 108 + extern int sysctl_llc2_rej_timeout; 109 + extern int sysctl_llc_station_ack_timeout; 110 + #else 111 + #define llc_sysctl_init() (0) 112 + #define llc_sysctl_exit() do { } while(0) 113 + #endif /* CONFIG_SYSCTL */ 118 114 #endif /* LLC_H */
+8 -7
include/net/llc_conn.h
··· 19 19 #define LLC_EVENT 1 20 20 #define LLC_PACKET 2 21 21 22 - #define LLC_P_TIME 2 23 - #define LLC_ACK_TIME 1 24 - #define LLC_REJ_TIME 3 25 - #define LLC_BUSY_TIME 3 22 + #define LLC2_P_TIME 2 23 + #define LLC2_ACK_TIME 1 24 + #define LLC2_REJ_TIME 3 25 + #define LLC2_BUSY_TIME 3 26 26 27 27 struct llc_timer { 28 28 struct timer_list timer; 29 - u16 expire; /* timer expire time */ 29 + unsigned long expire; /* timer expire time */ 30 30 }; 31 31 32 32 struct llc_sock { ··· 38 38 struct llc_addr laddr; /* lsap/mac pair */ 39 39 struct llc_addr daddr; /* dsap/mac pair */ 40 40 struct net_device *dev; /* device to send to remote */ 41 + u32 copied_seq; /* head of yet unread data */ 41 42 u8 retry_count; /* number of retries */ 42 43 u8 ack_must_be_send; 43 44 u8 first_pdu_Ns; ··· 93 92 return skb->cb[sizeof(skb->cb) - 1]; 94 93 } 95 94 96 - extern struct sock *llc_sk_alloc(int family, int priority, struct proto *prot); 95 + extern struct sock *llc_sk_alloc(int family, unsigned int __nocast priority, 96 + struct proto *prot); 97 97 extern void llc_sk_free(struct sock *sk); 98 98 99 99 extern void llc_sk_reset(struct sock *sk); ··· 117 115 118 116 extern u8 llc_data_accept_state(u8 state); 119 117 extern void llc_build_offset_table(void); 120 - extern int llc_release_sockets(struct llc_sap *sap); 121 118 #endif /* LLC_CONN_H */
+6 -2
include/net/llc_sap.h
··· 12 12 * See the GNU General Public License for more details. 13 13 */ 14 14 struct llc_sap; 15 + struct net_device; 15 16 struct sk_buff; 17 + struct sock; 16 18 17 19 extern void llc_sap_rtn_pdu(struct llc_sap *sap, struct sk_buff *skb); 18 - extern void llc_save_primitive(struct sk_buff* skb, unsigned char prim); 19 - extern struct sk_buff *llc_alloc_frame(void); 20 + extern void llc_save_primitive(struct sock *sk, struct sk_buff* skb, 21 + unsigned char prim); 22 + extern struct sk_buff *llc_alloc_frame(struct sock *sk, 23 + struct net_device *dev); 20 24 21 25 extern void llc_build_and_send_test_pkt(struct llc_sap *sap, 22 26 struct sk_buff *skb,
+1 -1
net/802/p8022.c
··· 56 56 57 57 void unregister_8022_client(struct datalink_proto *proto) 58 58 { 59 - llc_sap_close(proto->sap); 59 + llc_sap_put(proto->sap); 60 60 kfree(proto); 61 61 } 62 62
+1 -1
net/802/psnap.c
··· 106 106 107 107 static void __exit snap_exit(void) 108 108 { 109 - llc_sap_close(snap_sap); 109 + llc_sap_put(snap_sap); 110 110 } 111 111 112 112 module_exit(snap_exit);
+1 -1
net/802/tr.c
··· 238 238 return trllc->ethertype; 239 239 } 240 240 241 - return ntohs(ETH_P_802_2); 241 + return ntohs(ETH_P_TR_802_2); 242 242 } 243 243 244 244 /*
+2
net/core/dev.c
··· 574 574 return dev; 575 575 } 576 576 577 + EXPORT_SYMBOL(dev_getbyhwaddr); 578 + 577 579 struct net_device *dev_getfirstbyhwtype(unsigned short type) 578 580 { 579 581 struct net_device *dev;
+3 -2
net/ipv4/netfilter/Makefile
··· 4 4 5 5 # objects for the standalone - connection tracking / NAT 6 6 ip_conntrack-objs := ip_conntrack_standalone.o ip_conntrack_core.o ip_conntrack_proto_generic.o ip_conntrack_proto_tcp.o ip_conntrack_proto_udp.o ip_conntrack_proto_icmp.o 7 - iptable_nat-objs := ip_nat_standalone.o ip_nat_rule.o ip_nat_core.o ip_nat_helper.o ip_nat_proto_unknown.o ip_nat_proto_tcp.o ip_nat_proto_udp.o ip_nat_proto_icmp.o 7 + ip_nat-objs := ip_nat_core.o ip_nat_helper.o ip_nat_proto_unknown.o ip_nat_proto_tcp.o ip_nat_proto_udp.o ip_nat_proto_icmp.o 8 + iptable_nat-objs := ip_nat_rule.o ip_nat_standalone.o 8 9 9 10 ip_conntrack_pptp-objs := ip_conntrack_helper_pptp.o ip_conntrack_proto_gre.o 10 11 ip_nat_pptp-objs := ip_nat_helper_pptp.o ip_nat_proto_gre.o ··· 41 40 # the three instances of ip_tables 42 41 obj-$(CONFIG_IP_NF_FILTER) += iptable_filter.o 43 42 obj-$(CONFIG_IP_NF_MANGLE) += iptable_mangle.o 44 - obj-$(CONFIG_IP_NF_NAT) += iptable_nat.o 43 + obj-$(CONFIG_IP_NF_NAT) += iptable_nat.o ip_nat.o 45 44 obj-$(CONFIG_IP_NF_RAW) += iptable_raw.o 46 45 47 46 # matches
+24 -11
net/ipv4/netfilter/ip_nat_core.c
··· 74 74 75 75 return p; 76 76 } 77 + EXPORT_SYMBOL_GPL(ip_nat_proto_find_get); 77 78 78 79 void 79 80 ip_nat_proto_put(struct ip_nat_protocol *p) 80 81 { 81 82 module_put(p->me); 82 83 } 84 + EXPORT_SYMBOL_GPL(ip_nat_proto_put); 83 85 84 86 /* We keep an extra hash for each conntrack, for fast searching. */ 85 87 static inline unsigned int ··· 113 111 return csum_fold(csum_partial((char *)diffs, sizeof(diffs), 114 112 oldcheck^0xFFFF)); 115 113 } 114 + EXPORT_SYMBOL(ip_nat_cheat_check); 116 115 117 116 /* Is this tuple already taken? (not by us) */ 118 117 int ··· 130 127 invert_tuplepr(&reply, tuple); 131 128 return ip_conntrack_tuple_taken(&reply, ignored_conntrack); 132 129 } 130 + EXPORT_SYMBOL(ip_nat_used_tuple); 133 131 134 132 /* If we source map this tuple so reply looks like reply_tuple, will 135 133 * that meet the constraints of range. */ ··· 351 347 352 348 return NF_ACCEPT; 353 349 } 350 + EXPORT_SYMBOL(ip_nat_setup_info); 354 351 355 352 /* Returns true if succeeded. */ 356 353 static int ··· 392 387 } 393 388 394 389 /* Do packet manipulations according to ip_nat_setup_info. */ 395 - unsigned int nat_packet(struct ip_conntrack *ct, 396 - enum ip_conntrack_info ctinfo, 397 - unsigned int hooknum, 398 - struct sk_buff **pskb) 390 + unsigned int ip_nat_packet(struct ip_conntrack *ct, 391 + enum ip_conntrack_info ctinfo, 392 + unsigned int hooknum, 393 + struct sk_buff **pskb) 399 394 { 400 395 enum ip_conntrack_dir dir = CTINFO2DIR(ctinfo); 401 396 unsigned long statusbit; ··· 422 417 } 423 418 return NF_ACCEPT; 424 419 } 420 + EXPORT_SYMBOL_GPL(ip_nat_packet); 425 421 426 422 /* Dir is direction ICMP is coming from (opposite to packet it contains) */ 427 - int icmp_reply_translation(struct sk_buff **pskb, 428 - struct ip_conntrack *ct, 429 - enum ip_nat_manip_type manip, 430 - enum ip_conntrack_dir dir) 423 + int ip_nat_icmp_reply_translation(struct sk_buff **pskb, 424 + struct ip_conntrack *ct, 425 + enum ip_nat_manip_type manip, 426 + enum ip_conntrack_dir dir) 431 427 { 432 428 struct { 433 429 struct icmphdr icmp; ··· 515 509 516 510 return 1; 517 511 } 512 + EXPORT_SYMBOL_GPL(ip_nat_icmp_reply_translation); 518 513 519 514 /* Protocol registration. */ 520 515 int ip_nat_protocol_register(struct ip_nat_protocol *proto) ··· 532 525 write_unlock_bh(&ip_nat_lock); 533 526 return ret; 534 527 } 528 + EXPORT_SYMBOL(ip_nat_protocol_register); 535 529 536 530 /* Noone stores the protocol anywhere; simply delete it. */ 537 531 void ip_nat_protocol_unregister(struct ip_nat_protocol *proto) ··· 544 536 /* Someone could be still looking at the proto in a bh. */ 545 537 synchronize_net(); 546 538 } 539 + EXPORT_SYMBOL(ip_nat_protocol_unregister); 547 540 548 541 #if defined(CONFIG_IP_NF_CONNTRACK_NETLINK) || \ 549 542 defined(CONFIG_IP_NF_CONNTRACK_NETLINK_MODULE) ··· 591 582 EXPORT_SYMBOL_GPL(ip_nat_port_range_to_nfattr); 592 583 #endif 593 584 594 - int __init ip_nat_init(void) 585 + static int __init ip_nat_init(void) 595 586 { 596 587 size_t i; 597 588 ··· 633 624 return 0; 634 625 } 635 626 636 - /* Not __exit: called from ip_nat_standalone.c:init_or_cleanup() --RR */ 637 - void ip_nat_cleanup(void) 627 + static void __exit ip_nat_cleanup(void) 638 628 { 639 629 ip_ct_iterate_cleanup(&clean_nat, NULL); 640 630 ip_conntrack_destroyed = NULL; 641 631 vfree(bysource); 642 632 } 633 + 634 + MODULE_LICENSE("GPL"); 635 + 636 + module_init(ip_nat_init); 637 + module_exit(ip_nat_cleanup);
+4
net/ipv4/netfilter/ip_nat_helper.c
··· 199 199 } 200 200 return 1; 201 201 } 202 + EXPORT_SYMBOL(ip_nat_mangle_tcp_packet); 202 203 203 204 /* Generic function for mangling variable-length address changes inside 204 205 * NATed UDP connections (like the CONNECT DATA XXXXX MESG XXXXX INDEX XXXXX ··· 257 256 258 257 return 1; 259 258 } 259 + EXPORT_SYMBOL(ip_nat_mangle_udp_packet); 260 260 261 261 /* Adjust one found SACK option including checksum correction */ 262 262 static void ··· 401 399 402 400 return 1; 403 401 } 402 + EXPORT_SYMBOL(ip_nat_seq_adjust); 404 403 405 404 /* Setup NAT on this expected conntrack so it follows master. */ 406 405 /* If we fail to get a free NAT slot, we'll get dropped on confirm */ ··· 428 425 /* hook doesn't matter, but it has to do destination manip */ 429 426 ip_nat_setup_info(ct, &range, NF_IP_PRE_ROUTING); 430 427 } 428 + EXPORT_SYMBOL(ip_nat_follow_master);
+4 -21
net/ipv4/netfilter/ip_nat_standalone.c
··· 108 108 case IP_CT_RELATED: 109 109 case IP_CT_RELATED+IP_CT_IS_REPLY: 110 110 if ((*pskb)->nh.iph->protocol == IPPROTO_ICMP) { 111 - if (!icmp_reply_translation(pskb, ct, maniptype, 112 - CTINFO2DIR(ctinfo))) 111 + if (!ip_nat_icmp_reply_translation(pskb, ct, maniptype, 112 + CTINFO2DIR(ctinfo))) 113 113 return NF_DROP; 114 114 else 115 115 return NF_ACCEPT; ··· 152 152 } 153 153 154 154 IP_NF_ASSERT(info); 155 - return nat_packet(ct, ctinfo, hooknum, pskb); 155 + return ip_nat_packet(ct, ctinfo, hooknum, pskb); 156 156 } 157 157 158 158 static unsigned int ··· 325 325 printk("ip_nat_init: can't setup rules.\n"); 326 326 goto cleanup_nothing; 327 327 } 328 - ret = ip_nat_init(); 329 - if (ret < 0) { 330 - printk("ip_nat_init: can't setup rules.\n"); 331 - goto cleanup_rule_init; 332 - } 333 328 ret = nf_register_hook(&ip_nat_in_ops); 334 329 if (ret < 0) { 335 330 printk("ip_nat_init: can't register in hook.\n"); 336 - goto cleanup_nat; 331 + goto cleanup_rule_init; 337 332 } 338 333 ret = nf_register_hook(&ip_nat_out_ops); 339 334 if (ret < 0) { ··· 369 374 nf_unregister_hook(&ip_nat_out_ops); 370 375 cleanup_inops: 371 376 nf_unregister_hook(&ip_nat_in_ops); 372 - cleanup_nat: 373 - ip_nat_cleanup(); 374 377 cleanup_rule_init: 375 378 ip_nat_rule_cleanup(); 376 379 cleanup_nothing: ··· 388 395 module_init(init); 389 396 module_exit(fini); 390 397 391 - EXPORT_SYMBOL(ip_nat_setup_info); 392 - EXPORT_SYMBOL(ip_nat_protocol_register); 393 - EXPORT_SYMBOL(ip_nat_protocol_unregister); 394 - EXPORT_SYMBOL_GPL(ip_nat_proto_find_get); 395 - EXPORT_SYMBOL_GPL(ip_nat_proto_put); 396 - EXPORT_SYMBOL(ip_nat_cheat_check); 397 - EXPORT_SYMBOL(ip_nat_mangle_tcp_packet); 398 - EXPORT_SYMBOL(ip_nat_mangle_udp_packet); 399 - EXPORT_SYMBOL(ip_nat_used_tuple); 400 - EXPORT_SYMBOL(ip_nat_follow_master); 401 398 MODULE_LICENSE("GPL");
+2
net/ipv6/addrconf.c
··· 3520 3520 if (err) 3521 3521 return err; 3522 3522 3523 + ip6_null_entry.rt6i_idev = in6_dev_get(&loopback_dev); 3524 + 3523 3525 register_netdevice_notifier(&ipv6_dev_notf); 3524 3526 3525 3527 #ifdef CONFIG_IPV6_PRIVACY
+1
net/llc/Makefile
··· 22 22 llc_sap.o llc_s_ac.o llc_s_ev.o llc_s_st.o af_llc.o llc_station.o 23 23 24 24 llc2-$(CONFIG_PROC_FS) += llc_proc.o 25 + llc2-$(CONFIG_SYSCTL) += sysctl_net_llc.o
+319 -216
net/llc/af_llc.c
··· 21 21 * See the GNU General Public License for more details. 22 22 */ 23 23 #include <linux/config.h> 24 + #include <linux/compiler.h> 24 25 #include <linux/kernel.h> 25 26 #include <linux/module.h> 26 27 #include <linux/rtnetlink.h> ··· 38 37 static struct sockaddr_llc llc_ui_addrnull; 39 38 static struct proto_ops llc_ui_ops; 40 39 41 - static int llc_ui_wait_for_conn(struct sock *sk, int timeout); 42 - static int llc_ui_wait_for_disc(struct sock *sk, int timeout); 43 - static int llc_ui_wait_for_data(struct sock *sk, int timeout); 44 - static int llc_ui_wait_for_busy_core(struct sock *sk, int timeout); 40 + static int llc_ui_wait_for_conn(struct sock *sk, long timeout); 41 + static int llc_ui_wait_for_disc(struct sock *sk, long timeout); 42 + static int llc_ui_wait_for_busy_core(struct sock *sk, long timeout); 45 43 46 44 #if 0 47 45 #define dprintk(args...) printk(KERN_DEBUG args) ··· 116 116 struct llc_sock* llc = llc_sk(sk); 117 117 int rc = 0; 118 118 119 - if (llc_data_accept_state(llc->state) || llc->p_flag) { 120 - int timeout = sock_sndtimeo(sk, noblock); 119 + if (unlikely(llc_data_accept_state(llc->state) || llc->p_flag)) { 120 + long timeout = sock_sndtimeo(sk, noblock); 121 121 122 122 rc = llc_ui_wait_for_busy_core(sk, timeout); 123 123 } 124 - if (!rc) 124 + if (unlikely(!rc)) 125 125 rc = llc_build_and_send_pkt(sk, skb); 126 126 return rc; 127 127 } ··· 155 155 struct sock *sk; 156 156 int rc = -ESOCKTNOSUPPORT; 157 157 158 - if (sock->type == SOCK_DGRAM || sock->type == SOCK_STREAM) { 158 + if (likely(sock->type == SOCK_DGRAM || sock->type == SOCK_STREAM)) { 159 159 rc = -ENOMEM; 160 160 sk = llc_sk_alloc(PF_LLC, GFP_KERNEL, &llc_proto); 161 161 if (sk) { ··· 177 177 struct sock *sk = sock->sk; 178 178 struct llc_sock *llc; 179 179 180 - if (!sk) 180 + if (unlikely(sk == NULL)) 181 181 goto out; 182 182 sock_hold(sk); 183 183 lock_sock(sk); ··· 189 189 if (!sock_flag(sk, SOCK_ZAPPED)) 190 190 llc_sap_remove_socket(llc->sap, sk); 191 191 release_sock(sk); 192 - if (llc->sap && hlist_empty(&llc->sap->sk_list.list)) { 193 - llc_release_sockets(llc->sap); 194 - llc_sap_close(llc->sap); 195 - } 196 192 if (llc->dev) 197 193 dev_put(llc->dev); 198 194 sock_put(sk); ··· 217 221 llc_ui_sap_last_autoport = i + 2; 218 222 goto out; 219 223 } 224 + llc_sap_put(sap); 220 225 } 221 226 llc_ui_sap_last_autoport = LLC_SAP_DYN_START; 222 227 tries++; ··· 228 231 } 229 232 230 233 /** 231 - * llc_ui_autobind - Bind a socket to a specific address. 232 - * @sk: Socket to bind an address to. 233 - * @addr: Address the user wants the socket bound to. 234 + * llc_ui_autobind - automatically bind a socket to a sap 235 + * @sock: socket to bind 236 + * @addr: address to connect to 234 237 * 235 - * Bind a socket to a specific address. For llc a user is able to bind to 236 - * a specific sap only or mac + sap. If the user only specifies a sap and 237 - * a null dmac (all zeros) the user is attempting to bind to an entire 238 - * sap. This will stop anyone else on the local system from using that 239 - * sap. If someone else has a mac + sap open the bind to null + sap will 240 - * fail. 241 - * If the user desires to bind to a specific mac + sap, it is possible to 242 - * have multiple sap connections via multiple macs. 243 - * Bind and autobind for that matter must enforce the correct sap usage 244 - * otherwise all hell will break loose. 238 + * Used by llc_ui_connect and llc_ui_sendmsg when the user hasn't 239 + * specifically used llc_ui_bind to bind to an specific address/sap 240 + * 245 241 * Returns: 0 upon success, negative otherwise. 246 242 */ 247 243 static int llc_ui_autobind(struct socket *sock, struct sockaddr_llc *addr) ··· 275 285 * @addrlen: Length of the uaddr structure. 276 286 * 277 287 * Bind a socket to a specific address. For llc a user is able to bind to 278 - * a specific sap only or mac + sap. If the user only specifies a sap and 279 - * a null dmac (all zeros) the user is attempting to bind to an entire 280 - * sap. This will stop anyone else on the local system from using that 281 - * sap. If someone else has a mac + sap open the bind to null + sap will 282 - * fail. 288 + * a specific sap only or mac + sap. 283 289 * If the user desires to bind to a specific mac + sap, it is possible to 284 290 * have multiple sap connections via multiple macs. 285 291 * Bind and autobind for that matter must enforce the correct sap usage ··· 291 305 int rc = -EINVAL; 292 306 293 307 dprintk("%s: binding %02X\n", __FUNCTION__, addr->sllc_sap); 294 - if (!sock_flag(sk, SOCK_ZAPPED) || addrlen != sizeof(*addr)) 308 + if (unlikely(!sock_flag(sk, SOCK_ZAPPED) || addrlen != sizeof(*addr))) 295 309 goto out; 296 310 rc = -EAFNOSUPPORT; 297 - if (addr->sllc_family != AF_LLC) 311 + if (unlikely(addr->sllc_family != AF_LLC)) 312 + goto out; 313 + rc = -ENODEV; 314 + rtnl_lock(); 315 + llc->dev = dev_getbyhwaddr(addr->sllc_arphrd, addr->sllc_mac); 316 + rtnl_unlock(); 317 + if (!llc->dev) 298 318 goto out; 299 319 if (!addr->sllc_sap) { 300 320 rc = -EUSERS; ··· 314 322 rc = -EBUSY; /* some other network layer is using the sap */ 315 323 if (!sap) 316 324 goto out; 325 + llc_sap_hold(sap); 317 326 } else { 318 327 struct llc_addr laddr, daddr; 319 328 struct sock *ask; ··· 331 338 ask = llc_lookup_established(sap, &daddr, &laddr); 332 339 if (ask) { 333 340 sock_put(ask); 334 - goto out; 341 + goto out_put; 335 342 } 336 343 } 337 344 llc->laddr.lsap = addr->sllc_sap; ··· 341 348 llc_sap_add_socket(sap, sk); 342 349 sock_reset_flag(sk, SOCK_ZAPPED); 343 350 rc = 0; 351 + out_put: 352 + llc_sap_put(sap); 344 353 out: 345 354 return rc; 346 355 } ··· 364 369 int rc = -ENOTCONN; 365 370 366 371 lock_sock(sk); 367 - if (sk->sk_state != TCP_ESTABLISHED) 372 + if (unlikely(sk->sk_state != TCP_ESTABLISHED)) 368 373 goto out; 369 374 rc = -EINVAL; 370 375 if (how != 2) ··· 399 404 struct sock *sk = sock->sk; 400 405 struct llc_sock *llc = llc_sk(sk); 401 406 struct sockaddr_llc *addr = (struct sockaddr_llc *)uaddr; 402 - struct net_device *dev; 403 407 int rc = -EINVAL; 404 408 405 409 lock_sock(sk); 406 - if (addrlen != sizeof(*addr)) 410 + if (unlikely(addrlen != sizeof(*addr))) 407 411 goto out; 408 412 rc = -EAFNOSUPPORT; 409 - if (addr->sllc_family != AF_LLC) 413 + if (unlikely(addr->sllc_family != AF_LLC)) 414 + goto out; 415 + if (unlikely(sk->sk_type != SOCK_STREAM)) 416 + goto out; 417 + rc = -EALREADY; 418 + if (unlikely(sock->state == SS_CONNECTING)) 410 419 goto out; 411 420 /* bind connection to sap if user hasn't done it. */ 412 421 if (sock_flag(sk, SOCK_ZAPPED)) { ··· 418 419 rc = llc_ui_autobind(sock, addr); 419 420 if (rc) 420 421 goto out; 421 - llc->daddr.lsap = addr->sllc_sap; 422 - memcpy(llc->daddr.mac, addr->sllc_mac, IFHWADDRLEN); 423 422 } 424 - dev = llc->dev; 425 - if (sk->sk_type != SOCK_STREAM) 426 - goto out; 427 - rc = -EALREADY; 428 - if (sock->state == SS_CONNECTING) 429 - goto out; 423 + llc->daddr.lsap = addr->sllc_sap; 424 + memcpy(llc->daddr.mac, addr->sllc_mac, IFHWADDRLEN); 430 425 sock->state = SS_CONNECTING; 431 426 sk->sk_state = TCP_SYN_SENT; 432 427 llc->link = llc_ui_next_link_no(llc->sap->laddr.lsap); 433 - rc = llc_establish_connection(sk, dev->dev_addr, 428 + rc = llc_establish_connection(sk, llc->dev->dev_addr, 434 429 addr->sllc_mac, addr->sllc_sap); 435 430 if (rc) { 436 431 dprintk("%s: llc_ui_send_conn failed :-(\n", __FUNCTION__); ··· 432 439 sk->sk_state = TCP_CLOSE; 433 440 goto out; 434 441 } 435 - rc = llc_ui_wait_for_conn(sk, sk->sk_rcvtimeo); 436 - if (rc) 437 - dprintk("%s: llc_ui_wait_for_conn failed=%d\n", __FUNCTION__, rc); 442 + 443 + if (sk->sk_state == TCP_SYN_SENT) { 444 + const long timeo = sock_sndtimeo(sk, flags & O_NONBLOCK); 445 + 446 + if (!timeo || !llc_ui_wait_for_conn(sk, timeo)) 447 + goto out; 448 + 449 + rc = sock_intr_errno(timeo); 450 + if (signal_pending(current)) 451 + goto out; 452 + } 453 + 454 + if (sk->sk_state == TCP_CLOSE) 455 + goto sock_error; 456 + 457 + sock->state = SS_CONNECTED; 458 + rc = 0; 438 459 out: 439 460 release_sock(sk); 440 461 return rc; 462 + sock_error: 463 + rc = sock_error(sk) ? : -ECONNABORTED; 464 + sock->state = SS_UNCONNECTED; 465 + goto out; 441 466 } 442 467 443 468 /** ··· 472 461 int rc = -EINVAL; 473 462 474 463 lock_sock(sk); 475 - if (sock->state != SS_UNCONNECTED) 464 + if (unlikely(sock->state != SS_UNCONNECTED)) 476 465 goto out; 477 466 rc = -EOPNOTSUPP; 478 - if (sk->sk_type != SOCK_STREAM) 467 + if (unlikely(sk->sk_type != SOCK_STREAM)) 479 468 goto out; 480 469 rc = -EAGAIN; 481 470 if (sock_flag(sk, SOCK_ZAPPED)) ··· 494 483 return rc; 495 484 } 496 485 497 - static int llc_ui_wait_for_disc(struct sock *sk, int timeout) 486 + static int llc_ui_wait_for_disc(struct sock *sk, long timeout) 498 487 { 499 - DECLARE_WAITQUEUE(wait, current); 500 - int rc; 501 - 502 - add_wait_queue_exclusive(sk->sk_sleep, &wait); 503 - for (;;) { 504 - __set_current_state(TASK_INTERRUPTIBLE); 505 - rc = 0; 506 - if (sk->sk_state != TCP_CLOSE) { 507 - release_sock(sk); 508 - timeout = schedule_timeout(timeout); 509 - lock_sock(sk); 510 - } else 511 - break; 512 - rc = -ERESTARTSYS; 513 - if (signal_pending(current)) 514 - break; 515 - rc = -EAGAIN; 516 - if (!timeout) 517 - break; 518 - } 519 - __set_current_state(TASK_RUNNING); 520 - remove_wait_queue(sk->sk_sleep, &wait); 521 - return rc; 522 - } 523 - 524 - static int llc_ui_wait_for_conn(struct sock *sk, int timeout) 525 - { 526 - DECLARE_WAITQUEUE(wait, current); 527 - int rc; 528 - 529 - add_wait_queue_exclusive(sk->sk_sleep, &wait); 530 - for (;;) { 531 - __set_current_state(TASK_INTERRUPTIBLE); 532 - rc = -EAGAIN; 533 - if (sk->sk_state == TCP_CLOSE) 534 - break; 535 - rc = 0; 536 - if (sk->sk_state != TCP_ESTABLISHED) { 537 - release_sock(sk); 538 - timeout = schedule_timeout(timeout); 539 - lock_sock(sk); 540 - } else 541 - break; 542 - rc = -ERESTARTSYS; 543 - if (signal_pending(current)) 544 - break; 545 - rc = -EAGAIN; 546 - if (!timeout) 547 - break; 548 - } 549 - __set_current_state(TASK_RUNNING); 550 - remove_wait_queue(sk->sk_sleep, &wait); 551 - return rc; 552 - } 553 - 554 - static int llc_ui_wait_for_data(struct sock *sk, int timeout) 555 - { 556 - DECLARE_WAITQUEUE(wait, current); 488 + DEFINE_WAIT(wait); 557 489 int rc = 0; 558 490 559 - add_wait_queue_exclusive(sk->sk_sleep, &wait); 560 - for (;;) { 561 - __set_current_state(TASK_INTERRUPTIBLE); 562 - if (sk->sk_shutdown & RCV_SHUTDOWN) 563 - break; 564 - /* 565 - * Well, if we have backlog, try to process it now. 566 - */ 567 - if (sk->sk_backlog.tail) { 568 - release_sock(sk); 569 - lock_sock(sk); 570 - } 571 - rc = 0; 572 - if (skb_queue_empty(&sk->sk_receive_queue)) { 573 - release_sock(sk); 574 - timeout = schedule_timeout(timeout); 575 - lock_sock(sk); 576 - } else 491 + while (1) { 492 + prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE); 493 + if (sk_wait_event(sk, &timeout, sk->sk_state == TCP_CLOSE)) 577 494 break; 578 495 rc = -ERESTARTSYS; 579 496 if (signal_pending(current)) ··· 509 570 rc = -EAGAIN; 510 571 if (!timeout) 511 572 break; 573 + rc = 0; 512 574 } 513 - __set_current_state(TASK_RUNNING); 514 - remove_wait_queue(sk->sk_sleep, &wait); 575 + finish_wait(sk->sk_sleep, &wait); 515 576 return rc; 516 577 } 517 578 518 - static int llc_ui_wait_for_busy_core(struct sock *sk, int timeout) 579 + static int llc_ui_wait_for_conn(struct sock *sk, long timeout) 519 580 { 520 - DECLARE_WAITQUEUE(wait, current); 581 + DEFINE_WAIT(wait); 582 + 583 + while (1) { 584 + prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE); 585 + if (sk_wait_event(sk, &timeout, sk->sk_state != TCP_SYN_SENT)) 586 + break; 587 + if (signal_pending(current) || !timeout) 588 + break; 589 + } 590 + finish_wait(sk->sk_sleep, &wait); 591 + return timeout; 592 + } 593 + 594 + static int llc_ui_wait_for_busy_core(struct sock *sk, long timeout) 595 + { 596 + DEFINE_WAIT(wait); 521 597 struct llc_sock *llc = llc_sk(sk); 522 598 int rc; 523 599 524 - add_wait_queue_exclusive(sk->sk_sleep, &wait); 525 - for (;;) { 526 - dprintk("%s: looping...\n", __FUNCTION__); 527 - __set_current_state(TASK_INTERRUPTIBLE); 528 - rc = -ENOTCONN; 529 - if (sk->sk_shutdown & RCV_SHUTDOWN) 530 - break; 600 + while (1) { 601 + prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE); 531 602 rc = 0; 532 - if (llc_data_accept_state(llc->state) || llc->p_flag) { 533 - release_sock(sk); 534 - timeout = schedule_timeout(timeout); 535 - lock_sock(sk); 536 - } else 603 + if (sk_wait_event(sk, &timeout, 604 + (sk->sk_shutdown & RCV_SHUTDOWN) || 605 + (!llc_data_accept_state(llc->state) && 606 + !llc->p_flag))) 537 607 break; 538 608 rc = -ERESTARTSYS; 539 609 if (signal_pending(current)) ··· 551 603 if (!timeout) 552 604 break; 553 605 } 554 - __set_current_state(TASK_RUNNING); 555 - remove_wait_queue(sk->sk_sleep, &wait); 606 + finish_wait(sk->sk_sleep, &wait); 607 + return rc; 608 + } 609 + 610 + static int llc_wait_data(struct sock *sk, long timeo) 611 + { 612 + int rc; 613 + 614 + while (1) { 615 + /* 616 + * POSIX 1003.1g mandates this order. 617 + */ 618 + if (sk->sk_err) { 619 + rc = sock_error(sk); 620 + break; 621 + } 622 + rc = 0; 623 + if (sk->sk_shutdown & RCV_SHUTDOWN) 624 + break; 625 + rc = -EAGAIN; 626 + if (!timeo) 627 + break; 628 + rc = sock_intr_errno(timeo); 629 + if (signal_pending(current)) 630 + break; 631 + rc = 0; 632 + if (sk_wait_data(sk, &timeo)) 633 + break; 634 + } 556 635 return rc; 557 636 } 558 637 ··· 602 627 dprintk("%s: accepting on %02X\n", __FUNCTION__, 603 628 llc_sk(sk)->laddr.lsap); 604 629 lock_sock(sk); 605 - if (sk->sk_type != SOCK_STREAM) 630 + if (unlikely(sk->sk_type != SOCK_STREAM)) 606 631 goto out; 607 632 rc = -EINVAL; 608 - if (sock->state != SS_UNCONNECTED || sk->sk_state != TCP_LISTEN) 633 + if (unlikely(sock->state != SS_UNCONNECTED || 634 + sk->sk_state != TCP_LISTEN)) 609 635 goto out; 610 636 /* wait for a connection to arrive. */ 611 - rc = llc_ui_wait_for_data(sk, sk->sk_rcvtimeo); 612 - if (rc) 613 - goto out; 637 + if (skb_queue_empty(&sk->sk_receive_queue)) { 638 + rc = llc_wait_data(sk, sk->sk_rcvtimeo); 639 + if (rc) 640 + goto out; 641 + } 614 642 dprintk("%s: got a new connection on %02X\n", __FUNCTION__, 615 643 llc_sk(sk)->laddr.lsap); 616 644 skb = skb_dequeue(&sk->sk_receive_queue); ··· 635 657 /* put original socket back into a clean listen state. */ 636 658 sk->sk_state = TCP_LISTEN; 637 659 sk->sk_ack_backlog--; 638 - skb->sk = NULL; 639 660 dprintk("%s: ok success on %02X, client on %02X\n", __FUNCTION__, 640 661 llc_sk(sk)->addr.sllc_sap, newllc->daddr.lsap); 641 662 frees: ··· 648 671 * llc_ui_recvmsg - copy received data to the socket user. 649 672 * @sock: Socket to copy data from. 650 673 * @msg: Various user space related information. 651 - * @size: Size of user buffer. 674 + * @len: Size of user buffer. 652 675 * @flags: User specified flags. 653 676 * 654 677 * Copy received data to the socket user. 655 678 * Returns non-negative upon success, negative otherwise. 656 679 */ 657 680 static int llc_ui_recvmsg(struct kiocb *iocb, struct socket *sock, 658 - struct msghdr *msg, size_t size, int flags) 681 + struct msghdr *msg, size_t len, int flags) 659 682 { 660 - struct sock *sk = sock->sk; 661 683 struct sockaddr_llc *uaddr = (struct sockaddr_llc *)msg->msg_name; 662 - struct sk_buff *skb; 684 + const int nonblock = flags & MSG_DONTWAIT; 685 + struct sk_buff *skb = NULL; 686 + struct sock *sk = sock->sk; 687 + struct llc_sock *llc = llc_sk(sk); 663 688 size_t copied = 0; 664 - int rc = -ENOMEM, timeout; 665 - int noblock = flags & MSG_DONTWAIT; 689 + u32 peek_seq = 0; 690 + u32 *seq; 691 + unsigned long used; 692 + int target; /* Read at least this many bytes */ 693 + long timeo; 666 694 667 - dprintk("%s: receiving in %02X from %02X\n", __FUNCTION__, 668 - llc_sk(sk)->laddr.lsap, llc_sk(sk)->daddr.lsap); 669 695 lock_sock(sk); 670 - timeout = sock_rcvtimeo(sk, noblock); 671 - rc = llc_ui_wait_for_data(sk, timeout); 672 - if (rc) { 673 - dprintk("%s: llc_ui_wait_for_data failed recv " 674 - "in %02X from %02X\n", __FUNCTION__, 675 - llc_sk(sk)->laddr.lsap, llc_sk(sk)->daddr.lsap); 696 + copied = -ENOTCONN; 697 + if (sk->sk_state == TCP_LISTEN) 676 698 goto out; 677 - } 678 - skb = skb_dequeue(&sk->sk_receive_queue); 679 - if (!skb) /* shutdown */ 680 - goto out; 681 - copied = skb->len; 682 - if (copied > size) 683 - copied = size; 684 - rc = skb_copy_datagram_iovec(skb, 0, msg->msg_iov, copied); 685 - if (rc) 686 - goto dgram_free; 687 - if (skb->len > copied) { 688 - skb_pull(skb, copied); 689 - skb_queue_head(&sk->sk_receive_queue, skb); 690 - } 691 - if (uaddr) 692 - memcpy(uaddr, llc_ui_skb_cb(skb), sizeof(*uaddr)); 693 - msg->msg_namelen = sizeof(*uaddr); 694 - if (!skb->next) { 695 - dgram_free: 696 - kfree_skb(skb); 697 - } 699 + 700 + timeo = sock_rcvtimeo(sk, nonblock); 701 + 702 + seq = &llc->copied_seq; 703 + if (flags & MSG_PEEK) { 704 + peek_seq = llc->copied_seq; 705 + seq = &peek_seq; 706 + } 707 + 708 + target = sock_rcvlowat(sk, flags & MSG_WAITALL, len); 709 + copied = 0; 710 + 711 + do { 712 + u32 offset; 713 + 714 + /* 715 + * We need to check signals first, to get correct SIGURG 716 + * handling. FIXME: Need to check this doesn't impact 1003.1g 717 + * and move it down to the bottom of the loop 718 + */ 719 + if (signal_pending(current)) { 720 + if (copied) 721 + break; 722 + copied = timeo ? sock_intr_errno(timeo) : -EAGAIN; 723 + break; 724 + } 725 + 726 + /* Next get a buffer. */ 727 + 728 + skb = skb_peek(&sk->sk_receive_queue); 729 + if (skb) { 730 + offset = *seq; 731 + goto found_ok_skb; 732 + } 733 + /* Well, if we have backlog, try to process it now yet. */ 734 + 735 + if (copied >= target && !sk->sk_backlog.tail) 736 + break; 737 + 738 + if (copied) { 739 + if (sk->sk_err || 740 + sk->sk_state == TCP_CLOSE || 741 + (sk->sk_shutdown & RCV_SHUTDOWN) || 742 + !timeo || 743 + (flags & MSG_PEEK)) 744 + break; 745 + } else { 746 + if (sock_flag(sk, SOCK_DONE)) 747 + break; 748 + 749 + if (sk->sk_err) { 750 + copied = sock_error(sk); 751 + break; 752 + } 753 + if (sk->sk_shutdown & RCV_SHUTDOWN) 754 + break; 755 + 756 + if (sk->sk_state == TCP_CLOSE) { 757 + if (!sock_flag(sk, SOCK_DONE)) { 758 + /* 759 + * This occurs when user tries to read 760 + * from never connected socket. 761 + */ 762 + copied = -ENOTCONN; 763 + break; 764 + } 765 + break; 766 + } 767 + if (!timeo) { 768 + copied = -EAGAIN; 769 + break; 770 + } 771 + } 772 + 773 + if (copied >= target) { /* Do not sleep, just process backlog. */ 774 + release_sock(sk); 775 + lock_sock(sk); 776 + } else 777 + sk_wait_data(sk, &timeo); 778 + 779 + if ((flags & MSG_PEEK) && peek_seq != llc->copied_seq) { 780 + if (net_ratelimit()) 781 + printk(KERN_DEBUG "LLC(%s:%d): Application " 782 + "bug, race in MSG_PEEK.\n", 783 + current->comm, current->pid); 784 + peek_seq = llc->copied_seq; 785 + } 786 + continue; 787 + found_ok_skb: 788 + /* Ok so how much can we use? */ 789 + used = skb->len - offset; 790 + if (len < used) 791 + used = len; 792 + 793 + if (!(flags & MSG_TRUNC)) { 794 + int rc = skb_copy_datagram_iovec(skb, offset, 795 + msg->msg_iov, used); 796 + if (rc) { 797 + /* Exception. Bailout! */ 798 + if (!copied) 799 + copied = -EFAULT; 800 + break; 801 + } 802 + } 803 + 804 + *seq += used; 805 + copied += used; 806 + len -= used; 807 + 808 + if (used + offset < skb->len) 809 + continue; 810 + 811 + if (!(flags & MSG_PEEK)) { 812 + sk_eat_skb(sk, skb); 813 + *seq = 0; 814 + } 815 + } while (len > 0); 816 + 817 + /* 818 + * According to UNIX98, msg_name/msg_namelen are ignored 819 + * on connected socket. -ANK 820 + * But... af_llc still doesn't have separate sets of methods for 821 + * SOCK_DGRAM and SOCK_STREAM :-( So we have to do this test, will 822 + * eventually fix this tho :-) -acme 823 + */ 824 + if (sk->sk_type == SOCK_DGRAM) 825 + goto copy_uaddr; 698 826 out: 699 827 release_sock(sk); 700 - return rc ? : copied; 828 + return copied; 829 + copy_uaddr: 830 + if (uaddr != NULL && skb != NULL) { 831 + memcpy(uaddr, llc_ui_skb_cb(skb), sizeof(*uaddr)); 832 + msg->msg_namelen = sizeof(*uaddr); 833 + } 834 + goto out; 701 835 } 702 836 703 837 /** ··· 828 740 struct sockaddr_llc *addr = (struct sockaddr_llc *)msg->msg_name; 829 741 int flags = msg->msg_flags; 830 742 int noblock = flags & MSG_DONTWAIT; 831 - struct net_device *dev; 832 743 struct sk_buff *skb; 833 744 size_t size = 0; 834 745 int rc = -EINVAL, copied = 0, hdrlen; ··· 850 763 if (rc) 851 764 goto release; 852 765 } 853 - dev = llc->dev; 854 - hdrlen = dev->hard_header_len + llc_ui_header_len(sk, addr); 766 + hdrlen = llc->dev->hard_header_len + llc_ui_header_len(sk, addr); 855 767 size = hdrlen + len; 856 - if (size > dev->mtu) 857 - size = dev->mtu; 768 + if (size > llc->dev->mtu) 769 + size = llc->dev->mtu; 858 770 copied = size - hdrlen; 859 771 release_sock(sk); 860 772 skb = sock_alloc_send_skb(sk, size, noblock, &rc); 861 773 lock_sock(sk); 862 774 if (!skb) 863 775 goto release; 864 - skb->sk = sk; 865 - skb->dev = dev; 776 + skb->dev = llc->dev; 866 777 skb->protocol = llc_proto_type(addr->sllc_arphrd); 867 778 skb_reserve(skb, hdrlen); 868 779 rc = memcpy_fromiovec(skb_put(skb, copied), msg->msg_iov, copied); ··· 885 800 if (!(sk->sk_type == SOCK_STREAM && !addr->sllc_ua)) 886 801 goto out; 887 802 rc = llc_ui_send_data(sk, skb, noblock); 888 - if (rc) 889 - dprintk("%s: llc_ui_send_data failed: %d\n", __FUNCTION__, rc); 890 803 out: 891 - if (rc) 804 + if (rc) { 892 805 kfree_skb(skb); 893 806 release: 894 - if (rc) 895 807 dprintk("%s: failed sending from %02X to %02X: %d\n", 896 808 __FUNCTION__, llc->laddr.lsap, llc->daddr.lsap, rc); 809 + } 897 810 release_sock(sk); 898 811 return rc ? : copied; 899 812 } ··· 978 895 int rc = -EINVAL, opt; 979 896 980 897 lock_sock(sk); 981 - if (level != SOL_LLC || optlen != sizeof(int)) 898 + if (unlikely(level != SOL_LLC || optlen != sizeof(int))) 982 899 goto out; 983 900 rc = get_user(opt, (int __user *)optval); 984 901 if (rc) ··· 998 915 case LLC_OPT_ACK_TMR_EXP: 999 916 if (opt > LLC_OPT_MAX_ACK_TMR_EXP) 1000 917 goto out; 1001 - llc->ack_timer.expire = opt; 918 + llc->ack_timer.expire = opt * HZ; 1002 919 break; 1003 920 case LLC_OPT_P_TMR_EXP: 1004 921 if (opt > LLC_OPT_MAX_P_TMR_EXP) 1005 922 goto out; 1006 - llc->pf_cycle_timer.expire = opt; 923 + llc->pf_cycle_timer.expire = opt * HZ; 1007 924 break; 1008 925 case LLC_OPT_REJ_TMR_EXP: 1009 926 if (opt > LLC_OPT_MAX_REJ_TMR_EXP) 1010 927 goto out; 1011 - llc->rej_sent_timer.expire = opt; 928 + llc->rej_sent_timer.expire = opt * HZ; 1012 929 break; 1013 930 case LLC_OPT_BUSY_TMR_EXP: 1014 931 if (opt > LLC_OPT_MAX_BUSY_TMR_EXP) 1015 932 goto out; 1016 - llc->busy_state_timer.expire = opt; 933 + llc->busy_state_timer.expire = opt * HZ; 1017 934 break; 1018 935 case LLC_OPT_TX_WIN: 1019 936 if (opt > LLC_OPT_MAX_WIN) ··· 1053 970 int val = 0, len = 0, rc = -EINVAL; 1054 971 1055 972 lock_sock(sk); 1056 - if (level != SOL_LLC) 973 + if (unlikely(level != SOL_LLC)) 1057 974 goto out; 1058 975 rc = get_user(len, optlen); 1059 976 if (rc) ··· 1063 980 goto out; 1064 981 switch (optname) { 1065 982 case LLC_OPT_RETRY: 1066 - val = llc->n2; break; 983 + val = llc->n2; break; 1067 984 case LLC_OPT_SIZE: 1068 - val = llc->n1; break; 985 + val = llc->n1; break; 1069 986 case LLC_OPT_ACK_TMR_EXP: 1070 - val = llc->ack_timer.expire; break; 987 + val = llc->ack_timer.expire / HZ; break; 1071 988 case LLC_OPT_P_TMR_EXP: 1072 - val = llc->pf_cycle_timer.expire; break; 989 + val = llc->pf_cycle_timer.expire / HZ; break; 1073 990 case LLC_OPT_REJ_TMR_EXP: 1074 - val = llc->rej_sent_timer.expire; break; 991 + val = llc->rej_sent_timer.expire / HZ; break; 1075 992 case LLC_OPT_BUSY_TMR_EXP: 1076 - val = llc->busy_state_timer.expire; break; 993 + val = llc->busy_state_timer.expire / HZ; break; 1077 994 case LLC_OPT_TX_WIN: 1078 995 val = llc->k; break; 1079 996 case LLC_OPT_RX_WIN: ··· 1117 1034 .sendpage = sock_no_sendpage, 1118 1035 }; 1119 1036 1120 - extern void llc_sap_handler(struct llc_sap *sap, struct sk_buff *skb); 1121 - extern void llc_conn_handler(struct llc_sap *sap, struct sk_buff *skb); 1037 + static char llc_proc_err_msg[] __initdata = 1038 + KERN_CRIT "LLC: Unable to register the proc_fs entries\n"; 1039 + static char llc_sysctl_err_msg[] __initdata = 1040 + KERN_CRIT "LLC: Unable to register the sysctl entries\n"; 1041 + static char llc_sock_err_msg[] __initdata = 1042 + KERN_CRIT "LLC: Unable to register the network family\n"; 1122 1043 1123 1044 static int __init llc2_init(void) 1124 1045 { ··· 1135 1048 llc_station_init(); 1136 1049 llc_ui_sap_last_autoport = LLC_SAP_DYN_START; 1137 1050 rc = llc_proc_init(); 1138 - if (rc != 0) 1051 + if (rc != 0) { 1052 + printk(llc_proc_err_msg); 1139 1053 goto out_unregister_llc_proto; 1140 - sock_register(&llc_ui_family_ops); 1054 + } 1055 + rc = llc_sysctl_init(); 1056 + if (rc) { 1057 + printk(llc_sysctl_err_msg); 1058 + goto out_proc; 1059 + } 1060 + rc = sock_register(&llc_ui_family_ops); 1061 + if (rc) { 1062 + printk(llc_sock_err_msg); 1063 + goto out_sysctl; 1064 + } 1141 1065 llc_add_pack(LLC_DEST_SAP, llc_sap_handler); 1142 1066 llc_add_pack(LLC_DEST_CONN, llc_conn_handler); 1143 1067 out: 1144 1068 return rc; 1069 + out_sysctl: 1070 + llc_sysctl_exit(); 1071 + out_proc: 1072 + llc_proc_exit(); 1145 1073 out_unregister_llc_proto: 1146 1074 proto_unregister(&llc_proto); 1147 1075 goto out; ··· 1169 1067 llc_remove_pack(LLC_DEST_CONN); 1170 1068 sock_unregister(PF_LLC); 1171 1069 llc_proc_exit(); 1070 + llc_sysctl_exit(); 1172 1071 proto_unregister(&llc_proto); 1173 1072 } 1174 1073
+98 -173
net/llc/llc_c_ac.c
··· 60 60 61 61 int llc_conn_ac_conn_ind(struct sock *sk, struct sk_buff *skb) 62 62 { 63 - int rc = -ENOTCONN; 64 - u8 dsap; 65 - struct llc_sap *sap; 63 + struct llc_conn_state_ev *ev = llc_conn_ev(skb); 66 64 67 - llc_pdu_decode_dsap(skb, &dsap); 68 - sap = llc_sap_find(dsap); 69 - if (sap) { 70 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 71 - struct llc_sock *llc = llc_sk(sk); 72 - 73 - llc_pdu_decode_sa(skb, llc->daddr.mac); 74 - llc_pdu_decode_da(skb, llc->laddr.mac); 75 - llc->dev = skb->dev; 76 - ev->ind_prim = LLC_CONN_PRIM; 77 - rc = 0; 78 - } 79 - return rc; 65 + ev->ind_prim = LLC_CONN_PRIM; 66 + return 0; 80 67 } 81 68 82 69 int llc_conn_ac_conn_confirm(struct sock *sk, struct sk_buff *skb) ··· 107 120 reason = LLC_DISC_REASON_RX_DISC_CMD_PDU; 108 121 } else if (ev->type == LLC_CONN_EV_TYPE_ACK_TMR) 109 122 reason = LLC_DISC_REASON_ACK_TMR_EXP; 110 - else { 111 - reason = 0; 123 + else 112 124 rc = -EINVAL; 113 - } 114 125 if (!rc) { 115 126 ev->reason = reason; 116 127 ev->ind_prim = LLC_DISC_PRIM; ··· 145 160 LLC_U_PDU_CMD(pdu) == LLC_2_PDU_CMD_SABME) { 146 161 reason = LLC_RESET_REASON_REMOTE; 147 162 rc = 0; 148 - } else { 149 - reason = 0; 150 - rc = 1; 151 163 } 152 164 break; 153 165 case LLC_CONN_EV_TYPE_ACK_TMR: ··· 154 172 if (llc->retry_count > llc->n2) { 155 173 reason = LLC_RESET_REASON_LOCAL; 156 174 rc = 0; 157 - } else 158 - rc = 1; 175 + } 159 176 break; 160 177 } 161 178 if (!rc) { ··· 198 217 int llc_conn_ac_send_disc_cmd_p_set_x(struct sock *sk, struct sk_buff *skb) 199 218 { 200 219 int rc = -ENOBUFS; 201 - struct sk_buff *nskb = llc_alloc_frame(); 220 + struct llc_sock *llc = llc_sk(sk); 221 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 202 222 203 223 if (nskb) { 204 - struct llc_sock *llc = llc_sk(sk); 205 224 struct llc_sap *sap = llc->sap; 206 225 207 - nskb->dev = llc->dev; 208 226 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 209 227 llc->daddr.lsap, LLC_PDU_CMD); 210 228 llc_pdu_init_as_disc_cmd(nskb, 1); 211 229 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 212 - if (rc) 230 + if (unlikely(rc)) 213 231 goto free; 214 232 llc_conn_send_pdu(sk, nskb); 215 233 llc_conn_ac_set_p_flag_1(sk, skb); ··· 223 243 int llc_conn_ac_send_dm_rsp_f_set_p(struct sock *sk, struct sk_buff *skb) 224 244 { 225 245 int rc = -ENOBUFS; 226 - struct sk_buff *nskb = llc_alloc_frame(); 246 + struct llc_sock *llc = llc_sk(sk); 247 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 227 248 228 249 if (nskb) { 229 - struct llc_sock *llc = llc_sk(sk); 230 250 struct llc_sap *sap = llc->sap; 231 251 u8 f_bit; 232 252 233 - nskb->dev = llc->dev; 234 253 llc_pdu_decode_pf_bit(skb, &f_bit); 235 254 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 236 255 llc->daddr.lsap, LLC_PDU_RSP); 237 256 llc_pdu_init_as_dm_rsp(nskb, f_bit); 238 257 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 239 - if (rc) 258 + if (unlikely(rc)) 240 259 goto free; 241 260 llc_conn_send_pdu(sk, nskb); 242 261 } ··· 249 270 int llc_conn_ac_send_dm_rsp_f_set_1(struct sock *sk, struct sk_buff *skb) 250 271 { 251 272 int rc = -ENOBUFS; 252 - struct sk_buff *nskb = llc_alloc_frame(); 273 + struct llc_sock *llc = llc_sk(sk); 274 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 253 275 254 276 if (nskb) { 255 - struct llc_sock *llc = llc_sk(sk); 256 277 struct llc_sap *sap = llc->sap; 257 - u8 f_bit = 1; 258 278 259 - nskb->dev = llc->dev; 260 279 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 261 280 llc->daddr.lsap, LLC_PDU_RSP); 262 - llc_pdu_init_as_dm_rsp(nskb, f_bit); 281 + llc_pdu_init_as_dm_rsp(nskb, 1); 263 282 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 264 - if (rc) 283 + if (unlikely(rc)) 265 284 goto free; 266 285 llc_conn_send_pdu(sk, nskb); 267 286 } ··· 283 306 llc_pdu_decode_pf_bit(skb, &f_bit); 284 307 else 285 308 f_bit = 0; 286 - nskb = llc_alloc_frame(); 309 + nskb = llc_alloc_frame(sk, llc->dev); 287 310 if (nskb) { 288 311 struct llc_sap *sap = llc->sap; 289 312 290 - nskb->dev = llc->dev; 291 313 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 292 314 llc->daddr.lsap, LLC_PDU_RSP); 293 315 llc_pdu_init_as_frmr_rsp(nskb, pdu, f_bit, llc->vS, 294 316 llc->vR, INCORRECT); 295 317 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 296 - if (rc) 318 + if (unlikely(rc)) 297 319 goto free; 298 320 llc_conn_send_pdu(sk, nskb); 299 321 } ··· 306 330 int llc_conn_ac_resend_frmr_rsp_f_set_0(struct sock *sk, struct sk_buff *skb) 307 331 { 308 332 int rc = -ENOBUFS; 309 - struct sk_buff *nskb = llc_alloc_frame(); 333 + struct llc_sock *llc = llc_sk(sk); 334 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 310 335 311 336 if (nskb) { 312 - u8 f_bit = 0; 313 - struct llc_sock *llc = llc_sk(sk); 314 337 struct llc_sap *sap = llc->sap; 315 338 struct llc_pdu_sn *pdu = (struct llc_pdu_sn *)&llc->rx_pdu_hdr; 316 339 317 - nskb->dev = llc->dev; 318 340 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 319 341 llc->daddr.lsap, LLC_PDU_RSP); 320 - llc_pdu_init_as_frmr_rsp(nskb, pdu, f_bit, llc->vS, 342 + llc_pdu_init_as_frmr_rsp(nskb, pdu, 0, llc->vS, 321 343 llc->vR, INCORRECT); 322 344 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 323 - if (rc) 345 + if (unlikely(rc)) 324 346 goto free; 325 347 llc_conn_send_pdu(sk, nskb); 326 348 } ··· 334 360 u8 f_bit; 335 361 int rc = -ENOBUFS; 336 362 struct sk_buff *nskb; 363 + struct llc_sock *llc = llc_sk(sk); 337 364 338 365 llc_pdu_decode_pf_bit(skb, &f_bit); 339 - nskb = llc_alloc_frame(); 366 + nskb = llc_alloc_frame(sk, llc->dev); 340 367 if (nskb) { 341 - struct llc_sock *llc = llc_sk(sk); 342 368 struct llc_sap *sap = llc->sap; 343 369 struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 344 370 345 - nskb->dev = llc->dev; 346 371 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 347 372 llc->daddr.lsap, LLC_PDU_RSP); 348 373 llc_pdu_init_as_frmr_rsp(nskb, pdu, f_bit, llc->vS, 349 374 llc->vR, INCORRECT); 350 375 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 351 - if (rc) 376 + if (unlikely(rc)) 352 377 goto free; 353 378 llc_conn_send_pdu(sk, nskb); 354 379 } ··· 368 395 llc->daddr.lsap, LLC_PDU_CMD); 369 396 llc_pdu_init_as_i_cmd(skb, 1, llc->vS, llc->vR); 370 397 rc = llc_mac_hdr_init(skb, llc->dev->dev_addr, llc->daddr.mac); 371 - if (!rc) { 398 + if (likely(!rc)) { 372 399 llc_conn_send_pdu(sk, skb); 373 400 llc_conn_ac_inc_vs_by_1(sk, skb); 374 401 } ··· 385 412 llc->daddr.lsap, LLC_PDU_CMD); 386 413 llc_pdu_init_as_i_cmd(skb, 0, llc->vS, llc->vR); 387 414 rc = llc_mac_hdr_init(skb, llc->dev->dev_addr, llc->daddr.mac); 388 - if (!rc) { 415 + if (likely(!rc)) { 389 416 llc_conn_send_pdu(sk, skb); 390 417 llc_conn_ac_inc_vs_by_1(sk, skb); 391 418 } ··· 402 429 llc->daddr.lsap, LLC_PDU_CMD); 403 430 llc_pdu_init_as_i_cmd(skb, 0, llc->vS, llc->vR); 404 431 rc = llc_mac_hdr_init(skb, llc->dev->dev_addr, llc->daddr.mac); 405 - if (!rc) { 432 + if (likely(!rc)) { 406 433 llc_conn_send_pdu(sk, skb); 407 434 llc_conn_ac_inc_vs_by_1(sk, skb); 408 435 } ··· 424 451 u8 nr; 425 452 struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 426 453 int rc = -ENOBUFS; 427 - struct sk_buff *nskb = llc_alloc_frame(); 454 + struct llc_sock *llc = llc_sk(sk); 455 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 428 456 429 457 if (nskb) { 430 - struct llc_sock *llc = llc_sk(sk); 431 458 struct llc_sap *sap = llc->sap; 432 459 433 - nskb->dev = llc->dev; 434 460 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 435 461 llc->daddr.lsap, LLC_PDU_RSP); 436 462 llc_pdu_init_as_rr_rsp(nskb, 0, llc->vR); 437 463 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 438 - if (!rc) 464 + if (likely(!rc)) 439 465 llc_conn_send_pdu(sk, nskb); 440 466 else 441 467 kfree_skb(skb); ··· 459 487 int llc_conn_ac_send_rej_cmd_p_set_1(struct sock *sk, struct sk_buff *skb) 460 488 { 461 489 int rc = -ENOBUFS; 462 - struct sk_buff *nskb = llc_alloc_frame(); 490 + struct llc_sock *llc = llc_sk(sk); 491 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 463 492 464 493 if (nskb) { 465 - struct llc_sock *llc = llc_sk(sk); 466 494 struct llc_sap *sap = llc->sap; 467 495 468 - nskb->dev = llc->dev; 469 496 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 470 497 llc->daddr.lsap, LLC_PDU_CMD); 471 498 llc_pdu_init_as_rej_cmd(nskb, 1, llc->vR); 472 499 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 473 - if (rc) 500 + if (unlikely(rc)) 474 501 goto free; 475 502 llc_conn_send_pdu(sk, nskb); 476 503 } ··· 483 512 int llc_conn_ac_send_rej_rsp_f_set_1(struct sock *sk, struct sk_buff *skb) 484 513 { 485 514 int rc = -ENOBUFS; 486 - struct sk_buff *nskb = llc_alloc_frame(); 515 + struct llc_sock *llc = llc_sk(sk); 516 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 487 517 488 518 if (nskb) { 489 - u8 f_bit = 1; 490 - struct llc_sock *llc = llc_sk(sk); 491 519 struct llc_sap *sap = llc->sap; 492 520 493 - nskb->dev = llc->dev; 494 521 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 495 522 llc->daddr.lsap, LLC_PDU_RSP); 496 - llc_pdu_init_as_rej_rsp(nskb, f_bit, llc->vR); 523 + llc_pdu_init_as_rej_rsp(nskb, 1, llc->vR); 497 524 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 498 - if (rc) 525 + if (unlikely(rc)) 499 526 goto free; 500 527 llc_conn_send_pdu(sk, nskb); 501 528 } ··· 507 538 int llc_conn_ac_send_rej_xxx_x_set_0(struct sock *sk, struct sk_buff *skb) 508 539 { 509 540 int rc = -ENOBUFS; 510 - struct sk_buff *nskb = llc_alloc_frame(); 541 + struct llc_sock *llc = llc_sk(sk); 542 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 511 543 512 544 if (nskb) { 513 - struct llc_sock *llc = llc_sk(sk); 514 545 struct llc_sap *sap = llc->sap; 515 - u8 f_bit = 0; 516 546 517 - nskb->dev = llc->dev; 518 547 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 519 548 llc->daddr.lsap, LLC_PDU_RSP); 520 - llc_pdu_init_as_rej_rsp(nskb, f_bit, llc->vR); 549 + llc_pdu_init_as_rej_rsp(nskb, 0, llc->vR); 521 550 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 522 - if (rc) 551 + if (unlikely(rc)) 523 552 goto free; 524 553 llc_conn_send_pdu(sk, nskb); 525 554 } ··· 531 564 int llc_conn_ac_send_rnr_cmd_p_set_1(struct sock *sk, struct sk_buff *skb) 532 565 { 533 566 int rc = -ENOBUFS; 534 - struct sk_buff *nskb = llc_alloc_frame(); 567 + struct llc_sock *llc = llc_sk(sk); 568 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 535 569 536 570 if (nskb) { 537 - struct llc_sock *llc = llc_sk(sk); 538 571 struct llc_sap *sap = llc->sap; 539 572 540 - nskb->dev = llc->dev; 541 573 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 542 574 llc->daddr.lsap, LLC_PDU_CMD); 543 575 llc_pdu_init_as_rnr_cmd(nskb, 1, llc->vR); 544 576 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 545 - if (rc) 577 + if (unlikely(rc)) 546 578 goto free; 547 579 llc_conn_send_pdu(sk, nskb); 548 580 } ··· 555 589 int llc_conn_ac_send_rnr_rsp_f_set_1(struct sock *sk, struct sk_buff *skb) 556 590 { 557 591 int rc = -ENOBUFS; 558 - struct sk_buff *nskb = llc_alloc_frame(); 592 + struct llc_sock *llc = llc_sk(sk); 593 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 559 594 560 595 if (nskb) { 561 - struct llc_sock *llc = llc_sk(sk); 562 596 struct llc_sap *sap = llc->sap; 563 - u8 f_bit = 1; 564 597 565 - nskb->dev = llc->dev; 566 598 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 567 599 llc->daddr.lsap, LLC_PDU_RSP); 568 - llc_pdu_init_as_rnr_rsp(nskb, f_bit, llc->vR); 600 + llc_pdu_init_as_rnr_rsp(nskb, 1, llc->vR); 569 601 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 570 - if (rc) 602 + if (unlikely(rc)) 571 603 goto free; 572 604 llc_conn_send_pdu(sk, nskb); 573 605 } ··· 579 615 int llc_conn_ac_send_rnr_xxx_x_set_0(struct sock *sk, struct sk_buff *skb) 580 616 { 581 617 int rc = -ENOBUFS; 582 - struct sk_buff *nskb = llc_alloc_frame(); 618 + struct llc_sock *llc = llc_sk(sk); 619 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 583 620 584 621 if (nskb) { 585 - u8 f_bit = 0; 586 - struct llc_sock *llc = llc_sk(sk); 587 622 struct llc_sap *sap = llc->sap; 588 623 589 - nskb->dev = llc->dev; 590 624 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 591 625 llc->daddr.lsap, LLC_PDU_RSP); 592 - llc_pdu_init_as_rnr_rsp(nskb, f_bit, llc->vR); 626 + llc_pdu_init_as_rnr_rsp(nskb, 0, llc->vR); 593 627 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 594 - if (rc) 628 + if (unlikely(rc)) 595 629 goto free; 596 630 llc_conn_send_pdu(sk, nskb); 597 631 } ··· 607 645 if (!llc->remote_busy_flag) { 608 646 llc->remote_busy_flag = 1; 609 647 mod_timer(&llc->busy_state_timer.timer, 610 - jiffies + llc->busy_state_timer.expire * HZ); 648 + jiffies + llc->busy_state_timer.expire); 611 649 } 612 650 return 0; 613 651 } ··· 615 653 int llc_conn_ac_opt_send_rnr_xxx_x_set_0(struct sock *sk, struct sk_buff *skb) 616 654 { 617 655 int rc = -ENOBUFS; 618 - struct sk_buff *nskb = llc_alloc_frame(); 656 + struct llc_sock *llc = llc_sk(sk); 657 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 619 658 620 659 if (nskb) { 621 - struct llc_sock *llc = llc_sk(sk); 622 660 struct llc_sap *sap = llc->sap; 623 661 624 - nskb->dev = llc->dev; 625 662 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 626 663 llc->daddr.lsap, LLC_PDU_RSP); 627 664 llc_pdu_init_as_rnr_rsp(nskb, 0, llc->vR); 628 665 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 629 - if (rc) 666 + if (unlikely(rc)) 630 667 goto free; 631 668 llc_conn_send_pdu(sk, nskb); 632 669 } ··· 639 678 int llc_conn_ac_send_rr_cmd_p_set_1(struct sock *sk, struct sk_buff *skb) 640 679 { 641 680 int rc = -ENOBUFS; 642 - struct sk_buff *nskb = llc_alloc_frame(); 681 + struct llc_sock *llc = llc_sk(sk); 682 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 643 683 644 684 if (nskb) { 645 - struct llc_sock *llc = llc_sk(sk); 646 685 struct llc_sap *sap = llc->sap; 647 686 648 - nskb->dev = llc->dev; 649 687 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 650 688 llc->daddr.lsap, LLC_PDU_CMD); 651 689 llc_pdu_init_as_rr_cmd(nskb, 1, llc->vR); 652 690 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 653 - if (rc) 691 + if (unlikely(rc)) 654 692 goto free; 655 693 llc_conn_send_pdu(sk, nskb); 656 694 } ··· 663 703 int llc_conn_ac_send_rr_rsp_f_set_1(struct sock *sk, struct sk_buff *skb) 664 704 { 665 705 int rc = -ENOBUFS; 666 - struct sk_buff *nskb = llc_alloc_frame(); 706 + struct llc_sock *llc = llc_sk(sk); 707 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 667 708 668 709 if (nskb) { 669 - struct llc_sock *llc = llc_sk(sk); 670 710 struct llc_sap *sap = llc->sap; 671 711 u8 f_bit = 1; 672 712 673 - nskb->dev = llc->dev; 674 713 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 675 714 llc->daddr.lsap, LLC_PDU_RSP); 676 715 llc_pdu_init_as_rr_rsp(nskb, f_bit, llc->vR); 677 716 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 678 - if (rc) 717 + if (unlikely(rc)) 679 718 goto free; 680 719 llc_conn_send_pdu(sk, nskb); 681 720 } ··· 688 729 int llc_conn_ac_send_ack_rsp_f_set_1(struct sock *sk, struct sk_buff *skb) 689 730 { 690 731 int rc = -ENOBUFS; 691 - struct sk_buff *nskb = llc_alloc_frame(); 732 + struct llc_sock *llc = llc_sk(sk); 733 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 692 734 693 735 if (nskb) { 694 - struct llc_sock *llc = llc_sk(sk); 695 736 struct llc_sap *sap = llc->sap; 696 - u8 f_bit = 1; 697 737 698 - nskb->dev = llc->dev; 699 738 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 700 739 llc->daddr.lsap, LLC_PDU_RSP); 701 - llc_pdu_init_as_rr_rsp(nskb, f_bit, llc->vR); 740 + llc_pdu_init_as_rr_rsp(nskb, 1, llc->vR); 702 741 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 703 - if (rc) 742 + if (unlikely(rc)) 704 743 goto free; 705 744 llc_conn_send_pdu(sk, nskb); 706 745 } ··· 712 755 int llc_conn_ac_send_rr_xxx_x_set_0(struct sock *sk, struct sk_buff *skb) 713 756 { 714 757 int rc = -ENOBUFS; 715 - struct sk_buff *nskb = llc_alloc_frame(); 758 + struct llc_sock *llc = llc_sk(sk); 759 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 716 760 717 761 if (nskb) { 718 - struct llc_sock *llc = llc_sk(sk); 719 762 struct llc_sap *sap = llc->sap; 720 763 721 - nskb->dev = llc->dev; 722 764 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 723 765 llc->daddr.lsap, LLC_PDU_RSP); 724 766 llc_pdu_init_as_rr_rsp(nskb, 0, llc->vR); 725 767 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 726 - if (rc) 768 + if (unlikely(rc)) 727 769 goto free; 728 770 llc_conn_send_pdu(sk, nskb); 729 771 } ··· 736 780 int llc_conn_ac_send_ack_xxx_x_set_0(struct sock *sk, struct sk_buff *skb) 737 781 { 738 782 int rc = -ENOBUFS; 739 - struct sk_buff *nskb = llc_alloc_frame(); 783 + struct llc_sock *llc = llc_sk(sk); 784 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 740 785 741 786 if (nskb) { 742 - struct llc_sock *llc = llc_sk(sk); 743 787 struct llc_sap *sap = llc->sap; 744 788 745 - nskb->dev = llc->dev; 746 789 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 747 790 llc->daddr.lsap, LLC_PDU_RSP); 748 791 llc_pdu_init_as_rr_rsp(nskb, 0, llc->vR); 749 792 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 750 - if (rc) 793 + if (unlikely(rc)) 751 794 goto free; 752 795 llc_conn_send_pdu(sk, nskb); 753 796 } ··· 770 815 int llc_conn_ac_send_sabme_cmd_p_set_x(struct sock *sk, struct sk_buff *skb) 771 816 { 772 817 int rc = -ENOBUFS; 773 - struct sk_buff *nskb = llc_alloc_frame(); 774 818 struct llc_sock *llc = llc_sk(sk); 819 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 775 820 776 821 if (nskb) { 777 822 struct llc_sap *sap = llc->sap; ··· 779 824 780 825 if (llc->dev->flags & IFF_LOOPBACK) 781 826 dmac = llc->dev->dev_addr; 782 - nskb->dev = llc->dev; 783 827 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, 784 828 llc->daddr.lsap, LLC_PDU_CMD); 785 829 llc_pdu_init_as_sabme_cmd(nskb, 1); 786 830 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, dmac); 787 - if (rc) 831 + if (unlikely(rc)) 788 832 goto free; 789 833 llc_conn_send_pdu(sk, nskb); 790 834 llc_conn_set_p_flag(sk, 1); ··· 799 845 { 800 846 u8 f_bit; 801 847 int rc = -ENOBUFS; 802 - struct sk_buff *nskb = llc_alloc_frame(); 848 + struct llc_sock *llc = llc_sk(sk); 849 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 803 850 804 851 llc_pdu_decode_pf_bit(skb, &f_bit); 805 852 if (nskb) { 806 - struct llc_sock *llc = llc_sk(sk); 807 853 struct llc_sap *sap = llc->sap; 808 854 809 855 nskb->dev = llc->dev; ··· 811 857 llc->daddr.lsap, LLC_PDU_RSP); 812 858 llc_pdu_init_as_ua_rsp(nskb, f_bit); 813 859 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 814 - if (rc) 860 + if (unlikely(rc)) 815 861 goto free; 816 862 llc_conn_send_pdu(sk, nskb); 817 863 } ··· 840 886 841 887 llc_conn_set_p_flag(sk, 1); 842 888 mod_timer(&llc->pf_cycle_timer.timer, 843 - jiffies + llc->pf_cycle_timer.expire * HZ); 889 + jiffies + llc->pf_cycle_timer.expire); 844 890 return 0; 845 891 } 846 892 ··· 911 957 llc->daddr.lsap, LLC_PDU_RSP); 912 958 llc_pdu_init_as_i_cmd(skb, llc->ack_pf, llc->vS, llc->vR); 913 959 rc = llc_mac_hdr_init(skb, llc->dev->dev_addr, llc->daddr.mac); 914 - if (!rc) { 960 + if (likely(!rc)) { 915 961 llc_conn_send_pdu(sk, skb); 916 962 llc_conn_ac_inc_vs_by_1(sk, skb); 917 963 } ··· 955 1001 struct sk_buff *skb) 956 1002 { 957 1003 int rc = -ENOBUFS; 958 - struct sk_buff *nskb = llc_alloc_frame(); 1004 + struct llc_sock *llc = llc_sk(sk); 1005 + struct sk_buff *nskb = llc_alloc_frame(sk, llc->dev); 959 1006 960 1007 if (nskb) { 961 - struct llc_sock *llc = llc_sk(sk); 962 1008 struct llc_sap *sap = llc->sap; 963 1009 964 - nskb->dev = llc->dev; 965 1010 llc_pdu_header_init(nskb, LLC_PDU_TYPE_S, sap->laddr.lsap, 966 1011 llc->daddr.lsap, LLC_PDU_RSP); 967 1012 llc_pdu_init_as_rr_rsp(nskb, llc->ack_pf, llc->vR); 968 1013 rc = llc_mac_hdr_init(nskb, llc->dev->dev_addr, llc->daddr.mac); 969 - if (rc) 1014 + if (unlikely(rc)) 970 1015 goto free; 971 1016 llc_conn_send_pdu(sk, nskb); 972 1017 } ··· 1118 1165 { 1119 1166 struct llc_sock *llc = llc_sk(sk); 1120 1167 1121 - mod_timer(&llc->ack_timer.timer, jiffies + llc->ack_timer.expire * HZ); 1168 + mod_timer(&llc->ack_timer.timer, jiffies + llc->ack_timer.expire); 1122 1169 return 0; 1123 1170 } 1124 1171 ··· 1127 1174 struct llc_sock *llc = llc_sk(sk); 1128 1175 1129 1176 mod_timer(&llc->rej_sent_timer.timer, 1130 - jiffies + llc->rej_sent_timer.expire * HZ); 1177 + jiffies + llc->rej_sent_timer.expire); 1131 1178 return 0; 1132 1179 } 1133 1180 ··· 1138 1185 1139 1186 if (!timer_pending(&llc->ack_timer.timer)) 1140 1187 mod_timer(&llc->ack_timer.timer, 1141 - jiffies + llc->ack_timer.expire * HZ); 1188 + jiffies + llc->ack_timer.expire); 1142 1189 return 0; 1143 1190 } 1144 1191 ··· 1186 1233 } 1187 1234 if (unacked) 1188 1235 mod_timer(&llc->ack_timer.timer, 1189 - jiffies + llc->ack_timer.expire * HZ); 1236 + jiffies + llc->ack_timer.expire); 1190 1237 } else if (llc->failed_data_req) { 1191 1238 u8 f_bit; 1192 1239 ··· 1307 1354 return 0; 1308 1355 } 1309 1356 1310 - int llc_conn_ac_inc_vs_by_1(struct sock *sk, struct sk_buff *skb) 1357 + static int llc_conn_ac_inc_vs_by_1(struct sock *sk, struct sk_buff *skb) 1311 1358 { 1312 1359 llc_sk(sk)->vS = (llc_sk(sk)->vS + 1) % 128; 1313 1360 return 0; 1314 1361 } 1315 1362 1316 - void llc_conn_pf_cycle_tmr_cb(unsigned long timeout_data) 1363 + static void llc_conn_tmr_common_cb(unsigned long timeout_data, u8 type) 1317 1364 { 1318 1365 struct sock *sk = (struct sock *)timeout_data; 1319 1366 struct sk_buff *skb = alloc_skb(0, GFP_ATOMIC); ··· 1322 1369 if (skb) { 1323 1370 struct llc_conn_state_ev *ev = llc_conn_ev(skb); 1324 1371 1325 - skb->sk = sk; 1326 - ev->type = LLC_CONN_EV_TYPE_P_TMR; 1372 + skb_set_owner_r(skb, sk); 1373 + ev->type = type; 1327 1374 llc_process_tmr_ev(sk, skb); 1328 1375 } 1329 1376 bh_unlock_sock(sk); 1377 + } 1378 + 1379 + void llc_conn_pf_cycle_tmr_cb(unsigned long timeout_data) 1380 + { 1381 + llc_conn_tmr_common_cb(timeout_data, LLC_CONN_EV_TYPE_P_TMR); 1330 1382 } 1331 1383 1332 1384 void llc_conn_busy_tmr_cb(unsigned long timeout_data) 1333 1385 { 1334 - struct sock *sk = (struct sock *)timeout_data; 1335 - struct sk_buff *skb = alloc_skb(0, GFP_ATOMIC); 1336 - 1337 - bh_lock_sock(sk); 1338 - if (skb) { 1339 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 1340 - 1341 - skb->sk = sk; 1342 - ev->type = LLC_CONN_EV_TYPE_BUSY_TMR; 1343 - llc_process_tmr_ev(sk, skb); 1344 - } 1345 - bh_unlock_sock(sk); 1386 + llc_conn_tmr_common_cb(timeout_data, LLC_CONN_EV_TYPE_BUSY_TMR); 1346 1387 } 1347 1388 1348 1389 void llc_conn_ack_tmr_cb(unsigned long timeout_data) 1349 1390 { 1350 - struct sock* sk = (struct sock *)timeout_data; 1351 - struct sk_buff *skb = alloc_skb(0, GFP_ATOMIC); 1352 - 1353 - bh_lock_sock(sk); 1354 - if (skb) { 1355 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 1356 - 1357 - skb->sk = sk; 1358 - ev->type = LLC_CONN_EV_TYPE_ACK_TMR; 1359 - llc_process_tmr_ev(sk, skb); 1360 - } 1361 - bh_unlock_sock(sk); 1391 + llc_conn_tmr_common_cb(timeout_data, LLC_CONN_EV_TYPE_ACK_TMR); 1362 1392 } 1363 1393 1364 1394 void llc_conn_rej_tmr_cb(unsigned long timeout_data) 1365 1395 { 1366 - struct sock *sk = (struct sock *)timeout_data; 1367 - struct sk_buff *skb = alloc_skb(0, GFP_ATOMIC); 1368 - 1369 - bh_lock_sock(sk); 1370 - if (skb) { 1371 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 1372 - 1373 - skb->sk = sk; 1374 - ev->type = LLC_CONN_EV_TYPE_REJ_TMR; 1375 - llc_process_tmr_ev(sk, skb); 1376 - } 1377 - bh_unlock_sock(sk); 1396 + llc_conn_tmr_common_cb(timeout_data, LLC_CONN_EV_TYPE_REJ_TMR); 1378 1397 } 1379 1398 1380 1399 int llc_conn_ac_rst_vs(struct sock *sk, struct sk_buff *skb)
+68 -89
net/llc/llc_c_ev.c
··· 37 37 #include <net/llc_conn.h> 38 38 #include <net/llc_sap.h> 39 39 #include <net/sock.h> 40 + #include <net/llc_c_ac.h> 40 41 #include <net/llc_c_ev.h> 41 42 #include <net/llc_pdu.h> 42 43 ··· 46 45 #else 47 46 #define dprintk(args...) 48 47 #endif 49 - 50 - extern u16 llc_circular_between(u8 a, u8 b, u8 c); 51 48 52 49 /** 53 50 * llc_util_ns_inside_rx_window - check if sequence number is in rx window ··· 98 99 99 100 int llc_conn_ev_conn_req(struct sock *sk, struct sk_buff *skb) 100 101 { 101 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 102 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 102 103 103 104 return ev->prim == LLC_CONN_PRIM && 104 105 ev->prim_type == LLC_PRIM_TYPE_REQ ? 0 : 1; ··· 106 107 107 108 int llc_conn_ev_data_req(struct sock *sk, struct sk_buff *skb) 108 109 { 109 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 110 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 110 111 111 112 return ev->prim == LLC_DATA_PRIM && 112 113 ev->prim_type == LLC_PRIM_TYPE_REQ ? 0 : 1; ··· 114 115 115 116 int llc_conn_ev_disc_req(struct sock *sk, struct sk_buff *skb) 116 117 { 117 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 118 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 118 119 119 120 return ev->prim == LLC_DISC_PRIM && 120 121 ev->prim_type == LLC_PRIM_TYPE_REQ ? 0 : 1; ··· 122 123 123 124 int llc_conn_ev_rst_req(struct sock *sk, struct sk_buff *skb) 124 125 { 125 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 126 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 126 127 127 128 return ev->prim == LLC_RESET_PRIM && 128 129 ev->prim_type == LLC_PRIM_TYPE_REQ ? 0 : 1; ··· 130 131 131 132 int llc_conn_ev_local_busy_detected(struct sock *sk, struct sk_buff *skb) 132 133 { 133 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 134 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 134 135 135 136 return ev->type == LLC_CONN_EV_TYPE_SIMPLE && 136 137 ev->prim_type == LLC_CONN_EV_LOCAL_BUSY_DETECTED ? 0 : 1; ··· 138 139 139 140 int llc_conn_ev_local_busy_cleared(struct sock *sk, struct sk_buff *skb) 140 141 { 141 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 142 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 142 143 143 144 return ev->type == LLC_CONN_EV_TYPE_SIMPLE && 144 145 ev->prim_type == LLC_CONN_EV_LOCAL_BUSY_CLEARED ? 0 : 1; ··· 151 152 152 153 int llc_conn_ev_rx_disc_cmd_pbit_set_x(struct sock *sk, struct sk_buff *skb) 153 154 { 154 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 155 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 155 156 156 157 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_U(pdu) && 157 158 LLC_U_PDU_CMD(pdu) == LLC_2_PDU_CMD_DISC ? 0 : 1; ··· 159 160 160 161 int llc_conn_ev_rx_dm_rsp_fbit_set_x(struct sock *sk, struct sk_buff *skb) 161 162 { 162 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 163 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 163 164 164 165 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_U(pdu) && 165 166 LLC_U_PDU_RSP(pdu) == LLC_2_PDU_RSP_DM ? 0 : 1; ··· 167 168 168 169 int llc_conn_ev_rx_frmr_rsp_fbit_set_x(struct sock *sk, struct sk_buff *skb) 169 170 { 170 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 171 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 171 172 172 173 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_U(pdu) && 173 174 LLC_U_PDU_RSP(pdu) == LLC_2_PDU_RSP_FRMR ? 0 : 1; ··· 175 176 176 177 int llc_conn_ev_rx_i_cmd_pbit_set_0(struct sock *sk, struct sk_buff *skb) 177 178 { 178 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 179 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 179 180 180 181 return llc_conn_space(sk, skb) && 181 182 LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ··· 185 186 186 187 int llc_conn_ev_rx_i_cmd_pbit_set_1(struct sock *sk, struct sk_buff *skb) 187 188 { 188 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 189 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 189 190 190 191 return llc_conn_space(sk, skb) && 191 192 LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ··· 196 197 int llc_conn_ev_rx_i_cmd_pbit_set_0_unexpd_ns(struct sock *sk, 197 198 struct sk_buff *skb) 198 199 { 199 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 200 - u8 vr = llc_sk(sk)->vR; 201 - u8 ns = LLC_I_GET_NS(pdu); 200 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 201 + const u8 vr = llc_sk(sk)->vR; 202 + const u8 ns = LLC_I_GET_NS(pdu); 202 203 203 204 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 204 205 LLC_I_PF_IS_0(pdu) && ns != vr && ··· 208 209 int llc_conn_ev_rx_i_cmd_pbit_set_1_unexpd_ns(struct sock *sk, 209 210 struct sk_buff *skb) 210 211 { 211 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 212 - u8 vr = llc_sk(sk)->vR; 213 - u8 ns = LLC_I_GET_NS(pdu); 212 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 213 + const u8 vr = llc_sk(sk)->vR; 214 + const u8 ns = LLC_I_GET_NS(pdu); 214 215 215 216 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 216 217 LLC_I_PF_IS_1(pdu) && ns != vr && ··· 220 221 int llc_conn_ev_rx_i_cmd_pbit_set_x_inval_ns(struct sock *sk, 221 222 struct sk_buff *skb) 222 223 { 223 - struct llc_pdu_sn * pdu = llc_pdu_sn_hdr(skb); 224 - u8 vr = llc_sk(sk)->vR; 225 - u8 ns = LLC_I_GET_NS(pdu); 226 - u16 rc = LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ns != vr && 224 + const struct llc_pdu_sn * pdu = llc_pdu_sn_hdr(skb); 225 + const u8 vr = llc_sk(sk)->vR; 226 + const u8 ns = LLC_I_GET_NS(pdu); 227 + const u16 rc = LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 228 + ns != vr && 227 229 llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; 228 230 if (!rc) 229 231 dprintk("%s: matched, state=%d, ns=%d, vr=%d\n", ··· 234 234 235 235 int llc_conn_ev_rx_i_rsp_fbit_set_0(struct sock *sk, struct sk_buff *skb) 236 236 { 237 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 237 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 238 238 239 239 return llc_conn_space(sk, skb) && 240 240 LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ··· 244 244 245 245 int llc_conn_ev_rx_i_rsp_fbit_set_1(struct sock *sk, struct sk_buff *skb) 246 246 { 247 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 247 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 248 248 249 249 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 250 250 LLC_I_PF_IS_1(pdu) && ··· 253 253 254 254 int llc_conn_ev_rx_i_rsp_fbit_set_x(struct sock *sk, struct sk_buff *skb) 255 255 { 256 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 256 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 257 257 258 258 return llc_conn_space(sk, skb) && 259 259 LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ··· 263 263 int llc_conn_ev_rx_i_rsp_fbit_set_0_unexpd_ns(struct sock *sk, 264 264 struct sk_buff *skb) 265 265 { 266 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 267 - u8 vr = llc_sk(sk)->vR; 268 - u8 ns = LLC_I_GET_NS(pdu); 266 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 267 + const u8 vr = llc_sk(sk)->vR; 268 + const u8 ns = LLC_I_GET_NS(pdu); 269 269 270 270 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 271 271 LLC_I_PF_IS_0(pdu) && ns != vr && ··· 275 275 int llc_conn_ev_rx_i_rsp_fbit_set_1_unexpd_ns(struct sock *sk, 276 276 struct sk_buff *skb) 277 277 { 278 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 279 - u8 vr = llc_sk(sk)->vR; 280 - u8 ns = LLC_I_GET_NS(pdu); 278 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 279 + const u8 vr = llc_sk(sk)->vR; 280 + const u8 ns = LLC_I_GET_NS(pdu); 281 281 282 282 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 283 283 LLC_I_PF_IS_1(pdu) && ns != vr && ··· 287 287 int llc_conn_ev_rx_i_rsp_fbit_set_x_unexpd_ns(struct sock *sk, 288 288 struct sk_buff *skb) 289 289 { 290 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 291 - u8 vr = llc_sk(sk)->vR; 292 - u8 ns = LLC_I_GET_NS(pdu); 290 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 291 + const u8 vr = llc_sk(sk)->vR; 292 + const u8 ns = LLC_I_GET_NS(pdu); 293 293 294 294 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ns != vr && 295 295 !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; ··· 298 298 int llc_conn_ev_rx_i_rsp_fbit_set_x_inval_ns(struct sock *sk, 299 299 struct sk_buff *skb) 300 300 { 301 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 302 - u8 vr = llc_sk(sk)->vR; 303 - u8 ns = LLC_I_GET_NS(pdu); 304 - u16 rc = LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && ns != vr && 301 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 302 + const u8 vr = llc_sk(sk)->vR; 303 + const u8 ns = LLC_I_GET_NS(pdu); 304 + const u16 rc = LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_I(pdu) && 305 + ns != vr && 305 306 llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; 306 307 if (!rc) 307 308 dprintk("%s: matched, state=%d, ns=%d, vr=%d\n", ··· 312 311 313 312 int llc_conn_ev_rx_rej_cmd_pbit_set_0(struct sock *sk, struct sk_buff *skb) 314 313 { 315 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 314 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 316 315 317 316 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 318 317 LLC_S_PF_IS_0(pdu) && ··· 321 320 322 321 int llc_conn_ev_rx_rej_cmd_pbit_set_1(struct sock *sk, struct sk_buff *skb) 323 322 { 324 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 323 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 325 324 326 325 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 327 326 LLC_S_PF_IS_1(pdu) && ··· 330 329 331 330 int llc_conn_ev_rx_rej_rsp_fbit_set_0(struct sock *sk, struct sk_buff *skb) 332 331 { 333 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 332 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 334 333 335 334 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 336 335 LLC_S_PF_IS_0(pdu) && ··· 339 338 340 339 int llc_conn_ev_rx_rej_rsp_fbit_set_1(struct sock *sk, struct sk_buff *skb) 341 340 { 342 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 341 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 343 342 344 343 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 345 344 LLC_S_PF_IS_1(pdu) && ··· 348 347 349 348 int llc_conn_ev_rx_rej_rsp_fbit_set_x(struct sock *sk, struct sk_buff *skb) 350 349 { 351 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 350 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 352 351 353 352 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 354 353 LLC_S_PDU_RSP(pdu) == LLC_2_PDU_RSP_REJ ? 0 : 1; ··· 356 355 357 356 int llc_conn_ev_rx_rnr_cmd_pbit_set_0(struct sock *sk, struct sk_buff *skb) 358 357 { 359 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 358 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 360 359 361 360 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 362 361 LLC_S_PF_IS_0(pdu) && ··· 365 364 366 365 int llc_conn_ev_rx_rnr_cmd_pbit_set_1(struct sock *sk, struct sk_buff *skb) 367 366 { 368 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 367 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 369 368 370 369 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 371 370 LLC_S_PF_IS_1(pdu) && ··· 374 373 375 374 int llc_conn_ev_rx_rnr_rsp_fbit_set_0(struct sock *sk, struct sk_buff *skb) 376 375 { 377 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 376 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 378 377 379 378 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 380 379 LLC_S_PF_IS_0(pdu) && ··· 383 382 384 383 int llc_conn_ev_rx_rnr_rsp_fbit_set_1(struct sock *sk, struct sk_buff *skb) 385 384 { 386 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 385 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 387 386 388 387 return LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 389 388 LLC_S_PF_IS_1(pdu) && ··· 392 391 393 392 int llc_conn_ev_rx_rr_cmd_pbit_set_0(struct sock *sk, struct sk_buff *skb) 394 393 { 395 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 394 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 396 395 397 396 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 398 397 LLC_S_PF_IS_0(pdu) && ··· 401 400 402 401 int llc_conn_ev_rx_rr_cmd_pbit_set_1(struct sock *sk, struct sk_buff *skb) 403 402 { 404 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 403 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 405 404 406 405 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_S(pdu) && 407 406 LLC_S_PF_IS_1(pdu) && ··· 410 409 411 410 int llc_conn_ev_rx_rr_rsp_fbit_set_0(struct sock *sk, struct sk_buff *skb) 412 411 { 413 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 412 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 414 413 415 414 return llc_conn_space(sk, skb) && 416 415 LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && ··· 420 419 421 420 int llc_conn_ev_rx_rr_rsp_fbit_set_1(struct sock *sk, struct sk_buff *skb) 422 421 { 423 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 422 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 424 423 425 424 return llc_conn_space(sk, skb) && 426 425 LLC_PDU_IS_RSP(pdu) && LLC_PDU_TYPE_IS_S(pdu) && ··· 430 429 431 430 int llc_conn_ev_rx_sabme_cmd_pbit_set_x(struct sock *sk, struct sk_buff *skb) 432 431 { 433 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 432 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 434 433 435 434 return LLC_PDU_IS_CMD(pdu) && LLC_PDU_TYPE_IS_U(pdu) && 436 435 LLC_U_PDU_CMD(pdu) == LLC_2_PDU_CMD_SABME ? 0 : 1; ··· 447 446 int llc_conn_ev_rx_xxx_cmd_pbit_set_1(struct sock *sk, struct sk_buff *skb) 448 447 { 449 448 u16 rc = 1; 450 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 449 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 451 450 452 451 if (LLC_PDU_IS_CMD(pdu)) { 453 452 if (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) { ··· 462 461 int llc_conn_ev_rx_xxx_cmd_pbit_set_x(struct sock *sk, struct sk_buff *skb) 463 462 { 464 463 u16 rc = 1; 465 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 464 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 466 465 467 466 if (LLC_PDU_IS_CMD(pdu)) { 468 467 if (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) ··· 478 477 return rc; 479 478 } 480 479 481 - int llc_conn_ev_rx_xxx_rsp_fbit_set_1(struct sock *sk, struct sk_buff *skb) 482 - { 483 - u16 rc = 1; 484 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 485 - 486 - if (LLC_PDU_IS_RSP(pdu)) { 487 - if (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) { 488 - if (LLC_I_PF_IS_1(pdu)) 489 - rc = 0; 490 - } else if (LLC_PDU_TYPE_IS_U(pdu)) 491 - switch (LLC_U_PDU_RSP(pdu)) { 492 - case LLC_2_PDU_RSP_UA: 493 - case LLC_2_PDU_RSP_DM: 494 - case LLC_2_PDU_RSP_FRMR: 495 - if (LLC_U_PF_IS_1(pdu)) 496 - rc = 0; 497 - break; 498 - } 499 - } 500 - return rc; 501 - } 502 - 503 480 int llc_conn_ev_rx_xxx_rsp_fbit_set_x(struct sock *sk, struct sk_buff *skb) 504 481 { 505 482 u16 rc = 1; 506 - struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 483 + const struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 507 484 508 485 if (LLC_PDU_IS_RSP(pdu)) { 509 486 if (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) ··· 503 524 struct sk_buff *skb) 504 525 { 505 526 u16 rc = 1; 506 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 507 - u8 vs = llc_sk(sk)->vS; 508 - u8 nr = LLC_I_GET_NR(pdu); 527 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 528 + const u8 vs = llc_sk(sk)->vS; 529 + const u8 nr = LLC_I_GET_NR(pdu); 509 530 510 531 if (LLC_PDU_IS_CMD(pdu) && 511 532 (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) && ··· 521 542 struct sk_buff *skb) 522 543 { 523 544 u16 rc = 1; 524 - struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 525 - u8 vs = llc_sk(sk)->vS; 526 - u8 nr = LLC_I_GET_NR(pdu); 545 + const struct llc_pdu_sn *pdu = llc_pdu_sn_hdr(skb); 546 + const u8 vs = llc_sk(sk)->vS; 547 + const u8 nr = LLC_I_GET_NR(pdu); 527 548 528 549 if (LLC_PDU_IS_RSP(pdu) && 529 550 (LLC_PDU_TYPE_IS_I(pdu) || LLC_PDU_TYPE_IS_S(pdu)) && ··· 542 563 543 564 int llc_conn_ev_p_tmr_exp(struct sock *sk, struct sk_buff *skb) 544 565 { 545 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 566 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 546 567 547 568 return ev->type != LLC_CONN_EV_TYPE_P_TMR; 548 569 } 549 570 550 571 int llc_conn_ev_ack_tmr_exp(struct sock *sk, struct sk_buff *skb) 551 572 { 552 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 573 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 553 574 554 575 return ev->type != LLC_CONN_EV_TYPE_ACK_TMR; 555 576 } 556 577 557 578 int llc_conn_ev_rej_tmr_exp(struct sock *sk, struct sk_buff *skb) 558 579 { 559 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 580 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 560 581 561 582 return ev->type != LLC_CONN_EV_TYPE_REJ_TMR; 562 583 } 563 584 564 585 int llc_conn_ev_busy_tmr_exp(struct sock *sk, struct sk_buff *skb) 565 586 { 566 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 587 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 567 588 568 589 return ev->type != LLC_CONN_EV_TYPE_BUSY_TMR; 569 590 } ··· 575 596 576 597 int llc_conn_ev_tx_buffer_full(struct sock *sk, struct sk_buff *skb) 577 598 { 578 - struct llc_conn_state_ev *ev = llc_conn_ev(skb); 599 + const struct llc_conn_state_ev *ev = llc_conn_ev(skb); 579 600 580 601 return ev->type == LLC_CONN_EV_TYPE_SIMPLE && 581 602 ev->prim_type == LLC_CONN_EV_TX_BUFF_FULL ? 0 : 1;
+124 -90
net/llc/llc_conn.c
··· 40 40 /* Offset table on connection states transition diagram */ 41 41 static int llc_offset_table[NBR_CONN_STATES][NBR_CONN_EV]; 42 42 43 + int sysctl_llc2_ack_timeout = LLC2_ACK_TIME * HZ; 44 + int sysctl_llc2_p_timeout = LLC2_P_TIME * HZ; 45 + int sysctl_llc2_rej_timeout = LLC2_REJ_TIME * HZ; 46 + int sysctl_llc2_busy_timeout = LLC2_BUSY_TIME * HZ; 47 + 43 48 /** 44 49 * llc_conn_state_process - sends event to connection state machine 45 50 * @sk: connection ··· 58 53 int llc_conn_state_process(struct sock *sk, struct sk_buff *skb) 59 54 { 60 55 int rc; 61 - struct llc_sock *llc = llc_sk(sk); 56 + struct llc_sock *llc = llc_sk(skb->sk); 62 57 struct llc_conn_state_ev *ev = llc_conn_ev(skb); 63 58 64 59 /* ··· 68 63 */ 69 64 skb_get(skb); 70 65 ev->ind_prim = ev->cfm_prim = 0; 71 - rc = llc_conn_service(sk, skb); /* sending event to state machine */ 72 - if (rc) { 66 + /* 67 + * Send event to state machine 68 + */ 69 + rc = llc_conn_service(skb->sk, skb); 70 + if (unlikely(rc != 0)) { 73 71 printk(KERN_ERR "%s: llc_conn_service failed\n", __FUNCTION__); 74 72 goto out_kfree_skb; 75 73 } 76 74 77 - if (!ev->ind_prim && !ev->cfm_prim) { 75 + if (unlikely(!ev->ind_prim && !ev->cfm_prim)) { 78 76 /* indicate or confirm not required */ 79 77 /* XXX this is not very pretty, perhaps we should store 80 78 * XXX indicate/confirm-needed state in the llc_conn_state_ev ··· 88 80 goto out_skb_put; 89 81 } 90 82 91 - if (ev->ind_prim && ev->cfm_prim) /* Paranoia */ 83 + if (unlikely(ev->ind_prim && ev->cfm_prim)) /* Paranoia */ 92 84 skb_get(skb); 93 85 94 86 switch (ev->ind_prim) { 95 87 case LLC_DATA_PRIM: 96 - llc_save_primitive(skb, LLC_DATA_PRIM); 97 - if (sock_queue_rcv_skb(sk, skb)) { 88 + llc_save_primitive(sk, skb, LLC_DATA_PRIM); 89 + if (unlikely(sock_queue_rcv_skb(sk, skb))) { 98 90 /* 99 91 * shouldn't happen 100 92 */ ··· 103 95 kfree_skb(skb); 104 96 } 105 97 break; 106 - case LLC_CONN_PRIM: { 107 - struct sock *parent = skb->sk; 108 - 109 - skb->sk = sk; 110 - skb_queue_tail(&parent->sk_receive_queue, skb); 111 - sk->sk_state_change(parent); 112 - } 98 + case LLC_CONN_PRIM: 99 + /* 100 + * Can't be sock_queue_rcv_skb, because we have to leave the 101 + * skb->sk pointing to the newly created struct sock in 102 + * llc_conn_handler. -acme 103 + */ 104 + skb_queue_tail(&sk->sk_receive_queue, skb); 105 + sk->sk_state_change(sk); 113 106 break; 114 107 case LLC_DISC_PRIM: 115 108 sock_hold(sk); ··· 120 111 sk->sk_socket->state = SS_UNCONNECTED; 121 112 sk->sk_state = TCP_CLOSE; 122 113 if (!sock_flag(sk, SOCK_DEAD)) { 123 - sk->sk_state_change(sk); 124 114 sock_set_flag(sk, SOCK_DEAD); 115 + sk->sk_state_change(sk); 125 116 } 126 117 } 127 118 kfree_skb(skb); ··· 474 465 } 475 466 476 467 /** 477 - * llc_lookup_established - Finds connection for the remote/local sap/mac 468 + * __llc_lookup_established - Finds connection for the remote/local sap/mac 478 469 * @sap: SAP 479 470 * @daddr: address of remote LLC (MAC + SAP) 480 471 * @laddr: address of local LLC (MAC + SAP) ··· 482 473 * Search connection list of the SAP and finds connection using the remote 483 474 * mac, remote sap, local mac, and local sap. Returns pointer for 484 475 * connection found, %NULL otherwise. 476 + * Caller has to make sure local_bh is disabled. 485 477 */ 486 - struct sock *llc_lookup_established(struct llc_sap *sap, struct llc_addr *daddr, 487 - struct llc_addr *laddr) 478 + static struct sock *__llc_lookup_established(struct llc_sap *sap, 479 + struct llc_addr *daddr, 480 + struct llc_addr *laddr) 488 481 { 489 482 struct sock *rc; 490 483 struct hlist_node *node; 491 484 492 - read_lock_bh(&sap->sk_list.lock); 485 + read_lock(&sap->sk_list.lock); 493 486 sk_for_each(rc, node, &sap->sk_list.list) { 494 487 struct llc_sock *llc = llc_sk(rc); 495 488 ··· 505 494 } 506 495 rc = NULL; 507 496 found: 508 - read_unlock_bh(&sap->sk_list.lock); 497 + read_unlock(&sap->sk_list.lock); 509 498 return rc; 499 + } 500 + 501 + struct sock *llc_lookup_established(struct llc_sap *sap, 502 + struct llc_addr *daddr, 503 + struct llc_addr *laddr) 504 + { 505 + struct sock *sk; 506 + 507 + local_bh_disable(); 508 + sk = __llc_lookup_established(sap, daddr, laddr); 509 + local_bh_enable(); 510 + return sk; 510 511 } 511 512 512 513 /** ··· 529 506 * Search connection list of the SAP and finds connection listening on 530 507 * local mac, and local sap. Returns pointer for parent socket found, 531 508 * %NULL otherwise. 509 + * Caller has to make sure local_bh is disabled. 532 510 */ 533 511 static struct sock *llc_lookup_listener(struct llc_sap *sap, 534 512 struct llc_addr *laddr) ··· 537 513 struct sock *rc; 538 514 struct hlist_node *node; 539 515 540 - read_lock_bh(&sap->sk_list.lock); 516 + read_lock(&sap->sk_list.lock); 541 517 sk_for_each(rc, node, &sap->sk_list.list) { 542 518 struct llc_sock *llc = llc_sk(rc); 543 519 ··· 551 527 } 552 528 rc = NULL; 553 529 found: 554 - read_unlock_bh(&sap->sk_list.lock); 530 + read_unlock(&sap->sk_list.lock); 555 531 return rc; 532 + } 533 + 534 + static struct sock *__llc_lookup(struct llc_sap *sap, 535 + struct llc_addr *daddr, 536 + struct llc_addr *laddr) 537 + { 538 + struct sock *sk = __llc_lookup_established(sap, daddr, laddr); 539 + 540 + return sk ? : llc_lookup_listener(sap, laddr); 556 541 } 557 542 558 543 /** ··· 577 544 } 578 545 579 546 /** 580 - * find_next_offset - finds offset for next category of transitions 547 + * llc_find_next_offset - finds offset for next category of transitions 581 548 * @state: state table. 582 549 * @offset: start offset. 583 550 * 584 551 * Finds offset of next category of transitions in transition table. 585 552 * Returns the start index of next category. 586 553 */ 587 - static u16 find_next_offset(struct llc_conn_state *state, u16 offset) 554 + static u16 __init llc_find_next_offset(struct llc_conn_state *state, u16 offset) 588 555 { 589 556 u16 cnt = 0; 590 557 struct llc_conn_state_trans **next_trans; ··· 611 578 next_offset = 0; 612 579 for (ev_type = 0; ev_type < NBR_CONN_EV; ev_type++) { 613 580 llc_offset_table[state][ev_type] = next_offset; 614 - next_offset += find_next_offset(curr_state, 615 - next_offset) + 1; 581 + next_offset += llc_find_next_offset(curr_state, 582 + next_offset) + 1; 616 583 } 617 584 } 618 585 } ··· 656 623 */ 657 624 void llc_sap_add_socket(struct llc_sap *sap, struct sock *sk) 658 625 { 626 + llc_sap_hold(sap); 659 627 write_lock_bh(&sap->sk_list.lock); 660 628 llc_sk(sk)->sap = sap; 661 629 sk_add_node(sk, &sap->sk_list.list); ··· 676 642 write_lock_bh(&sap->sk_list.lock); 677 643 sk_del_node_init(sk); 678 644 write_unlock_bh(&sap->sk_list.lock); 645 + llc_sap_put(sap); 679 646 } 680 647 681 648 /** ··· 689 654 static int llc_conn_rcv(struct sock* sk, struct sk_buff *skb) 690 655 { 691 656 struct llc_conn_state_ev *ev = llc_conn_ev(skb); 692 - struct llc_sock *llc = llc_sk(sk); 693 657 694 - if (!llc->dev) 695 - llc->dev = skb->dev; 696 658 ev->type = LLC_CONN_EV_TYPE_PDU; 697 659 ev->reason = 0; 698 660 return llc_conn_state_process(sk, skb); 661 + } 662 + 663 + static struct sock *llc_create_incoming_sock(struct sock *sk, 664 + struct net_device *dev, 665 + struct llc_addr *saddr, 666 + struct llc_addr *daddr) 667 + { 668 + struct sock *newsk = llc_sk_alloc(sk->sk_family, GFP_ATOMIC, 669 + sk->sk_prot); 670 + struct llc_sock *newllc, *llc = llc_sk(sk); 671 + 672 + if (!newsk) 673 + goto out; 674 + newllc = llc_sk(newsk); 675 + memcpy(&newllc->laddr, daddr, sizeof(newllc->laddr)); 676 + memcpy(&newllc->daddr, saddr, sizeof(newllc->daddr)); 677 + newllc->dev = dev; 678 + dev_hold(dev); 679 + llc_sap_add_socket(llc->sap, newsk); 680 + llc_sap_hold(llc->sap); 681 + out: 682 + return newsk; 699 683 } 700 684 701 685 void llc_conn_handler(struct llc_sap *sap, struct sk_buff *skb) ··· 727 673 llc_pdu_decode_da(skb, daddr.mac); 728 674 llc_pdu_decode_dsap(skb, &daddr.lsap); 729 675 730 - sk = llc_lookup_established(sap, &saddr, &daddr); 731 - if (!sk) { 732 - /* 733 - * Didn't find an active connection; verify if there 734 - * is a listening socket for this llc addr 735 - */ 736 - struct llc_sock *llc; 737 - struct sock *parent = llc_lookup_listener(sap, &daddr); 676 + sk = __llc_lookup(sap, &saddr, &daddr); 677 + if (!sk) 678 + goto drop; 738 679 739 - if (!parent) { 740 - dprintk("llc_lookup_listener failed!\n"); 741 - goto drop; 742 - } 743 - 744 - sk = llc_sk_alloc(parent->sk_family, GFP_ATOMIC, parent->sk_prot); 745 - if (!sk) { 746 - sock_put(parent); 747 - goto drop; 748 - } 749 - llc = llc_sk(sk); 750 - memcpy(&llc->laddr, &daddr, sizeof(llc->laddr)); 751 - memcpy(&llc->daddr, &saddr, sizeof(llc->daddr)); 752 - llc_sap_add_socket(sap, sk); 753 - sock_hold(sk); 754 - sock_put(parent); 755 - skb->sk = parent; 756 - } else 757 - skb->sk = sk; 758 680 bh_lock_sock(sk); 681 + /* 682 + * This has to be done here and not at the upper layer ->accept 683 + * method because of the way the PROCOM state machine works: 684 + * it needs to set several state variables (see, for instance, 685 + * llc_adm_actions_2 in net/llc/llc_c_st.c) and send a packet to 686 + * the originator of the new connection, and this state has to be 687 + * in the newly created struct sock private area. -acme 688 + */ 689 + if (unlikely(sk->sk_state == TCP_LISTEN)) { 690 + struct sock *newsk = llc_create_incoming_sock(sk, skb->dev, 691 + &saddr, &daddr); 692 + if (!newsk) 693 + goto drop_unlock; 694 + skb_set_owner_r(skb, newsk); 695 + } else { 696 + /* 697 + * Can't be skb_set_owner_r, this will be done at the 698 + * llc_conn_state_process function, later on, when we will use 699 + * skb_queue_rcv_skb to send it to upper layers, this is 700 + * another trick required to cope with how the PROCOM state 701 + * machine works. -acme 702 + */ 703 + skb->sk = sk; 704 + } 759 705 if (!sock_owned_by_user(sk)) 760 706 llc_conn_rcv(sk, skb); 761 707 else { ··· 763 709 llc_set_backlog_type(skb, LLC_PACKET); 764 710 sk_add_backlog(sk, skb); 765 711 } 712 + out: 766 713 bh_unlock_sock(sk); 767 714 sock_put(sk); 768 715 return; 769 716 drop: 770 717 kfree_skb(skb); 718 + return; 719 + drop_unlock: 720 + kfree_skb(skb); 721 + goto out; 771 722 } 772 723 773 724 #undef LLC_REFCNT_DEBUG 774 725 #ifdef LLC_REFCNT_DEBUG 775 726 static atomic_t llc_sock_nr; 776 727 #endif 777 - 778 - /** 779 - * llc_release_sockets - releases all sockets in a sap 780 - * @sap: sap to release its sockets 781 - * 782 - * Releases all connections of a sap. Returns 0 if all actions complete 783 - * successfully, nonzero otherwise 784 - */ 785 - int llc_release_sockets(struct llc_sap *sap) 786 - { 787 - int rc = 0; 788 - struct sock *sk; 789 - struct hlist_node *node; 790 - 791 - write_lock_bh(&sap->sk_list.lock); 792 - 793 - sk_for_each(sk, node, &sap->sk_list.list) { 794 - llc_sk(sk)->state = LLC_CONN_STATE_TEMP; 795 - 796 - if (llc_send_disc(sk)) 797 - rc = 1; 798 - } 799 - 800 - write_unlock_bh(&sap->sk_list.lock); 801 - return rc; 802 - } 803 728 804 729 /** 805 730 * llc_backlog_rcv - Processes rx frames and expired timers. ··· 795 762 int rc = 0; 796 763 struct llc_sock *llc = llc_sk(sk); 797 764 798 - if (llc_backlog_type(skb) == LLC_PACKET) { 799 - if (llc->state > 1) /* not closed */ 765 + if (likely(llc_backlog_type(skb) == LLC_PACKET)) { 766 + if (likely(llc->state > 1)) /* not closed */ 800 767 rc = llc_conn_rcv(sk, skb); 801 768 else 802 769 goto out_kfree_skb; 803 770 } else if (llc_backlog_type(skb) == LLC_EVENT) { 804 771 /* timer expiration event */ 805 - if (llc->state > 1) /* not closed */ 772 + if (likely(llc->state > 1)) /* not closed */ 806 773 rc = llc_conn_state_process(sk, skb); 807 774 else 808 775 goto out_kfree_skb; ··· 832 799 llc->dec_step = llc->connect_step = 1; 833 800 834 801 init_timer(&llc->ack_timer.timer); 835 - llc->ack_timer.expire = LLC_ACK_TIME; 802 + llc->ack_timer.expire = sysctl_llc2_ack_timeout; 836 803 llc->ack_timer.timer.data = (unsigned long)sk; 837 804 llc->ack_timer.timer.function = llc_conn_ack_tmr_cb; 838 805 839 806 init_timer(&llc->pf_cycle_timer.timer); 840 - llc->pf_cycle_timer.expire = LLC_P_TIME; 807 + llc->pf_cycle_timer.expire = sysctl_llc2_p_timeout; 841 808 llc->pf_cycle_timer.timer.data = (unsigned long)sk; 842 809 llc->pf_cycle_timer.timer.function = llc_conn_pf_cycle_tmr_cb; 843 810 844 811 init_timer(&llc->rej_sent_timer.timer); 845 - llc->rej_sent_timer.expire = LLC_REJ_TIME; 812 + llc->rej_sent_timer.expire = sysctl_llc2_rej_timeout; 846 813 llc->rej_sent_timer.timer.data = (unsigned long)sk; 847 814 llc->rej_sent_timer.timer.function = llc_conn_rej_tmr_cb; 848 815 849 816 init_timer(&llc->busy_state_timer.timer); 850 - llc->busy_state_timer.expire = LLC_BUSY_TIME; 817 + llc->busy_state_timer.expire = sysctl_llc2_busy_timeout; 851 818 llc->busy_state_timer.timer.data = (unsigned long)sk; 852 819 llc->busy_state_timer.timer.function = llc_conn_busy_tmr_cb; 853 820 ··· 867 834 * Allocates a LLC sock and initializes it. Returns the new LLC sock 868 835 * or %NULL if there's no memory available for one 869 836 */ 870 - struct sock *llc_sk_alloc(int family, int priority, struct proto *prot) 837 + struct sock *llc_sk_alloc(int family, unsigned int __nocast priority, 838 + struct proto *prot) 871 839 { 872 840 struct sock *sk = sk_alloc(family, priority, prot, 1); 873 841
+23 -11
net/llc/llc_core.c
··· 40 40 sap->state = LLC_SAP_STATE_ACTIVE; 41 41 memcpy(sap->laddr.mac, llc_station_mac_sa, ETH_ALEN); 42 42 rwlock_init(&sap->sk_list.lock); 43 + atomic_set(&sap->refcnt, 1); 43 44 } 44 45 return sap; 45 46 } ··· 53 52 */ 54 53 static void llc_add_sap(struct llc_sap *sap) 55 54 { 56 - write_lock_bh(&llc_sap_list_lock); 57 55 list_add_tail(&sap->node, &llc_sap_list); 58 - write_unlock_bh(&llc_sap_list_lock); 59 56 } 60 57 61 58 /** ··· 69 70 write_unlock_bh(&llc_sap_list_lock); 70 71 } 71 72 73 + static struct llc_sap *__llc_sap_find(unsigned char sap_value) 74 + { 75 + struct llc_sap* sap; 76 + 77 + list_for_each_entry(sap, &llc_sap_list, node) 78 + if (sap->laddr.lsap == sap_value) 79 + goto out; 80 + sap = NULL; 81 + out: 82 + return sap; 83 + } 84 + 72 85 /** 73 86 * llc_sap_find - searchs a SAP in station 74 87 * @sap_value: sap to be found 75 88 * 76 89 * Searchs for a sap in the sap list of the LLC's station upon the sap ID. 90 + * If the sap is found it will be refcounted and the user will have to do 91 + * a llc_sap_put after use. 77 92 * Returns the sap or %NULL if not found. 78 93 */ 79 94 struct llc_sap *llc_sap_find(unsigned char sap_value) ··· 95 82 struct llc_sap* sap; 96 83 97 84 read_lock_bh(&llc_sap_list_lock); 98 - list_for_each_entry(sap, &llc_sap_list, node) 99 - if (sap->laddr.lsap == sap_value) 100 - goto out; 101 - sap = NULL; 102 - out: 85 + sap = __llc_sap_find(sap_value); 86 + if (sap) 87 + llc_sap_hold(sap); 103 88 read_unlock_bh(&llc_sap_list_lock); 104 89 return sap; 105 90 } ··· 117 106 struct packet_type *pt, 118 107 struct net_device *orig_dev)) 119 108 { 120 - struct llc_sap *sap = llc_sap_find(lsap); 109 + struct llc_sap *sap = NULL; 121 110 122 - if (sap) { /* SAP already exists */ 123 - sap = NULL; 111 + write_lock_bh(&llc_sap_list_lock); 112 + if (__llc_sap_find(lsap)) /* SAP already exists */ 124 113 goto out; 125 - } 126 114 sap = llc_sap_alloc(); 127 115 if (!sap) 128 116 goto out; 129 117 sap->laddr.lsap = lsap; 130 118 sap->rcv_func = func; 119 + llc_sap_hold(sap); 131 120 llc_add_sap(sap); 132 121 out: 122 + write_unlock_bh(&llc_sap_list_lock); 133 123 return sap; 134 124 } 135 125
+5 -6
net/llc/llc_if.c
··· 47 47 int rc = -ECONNABORTED; 48 48 struct llc_sock *llc = llc_sk(sk); 49 49 50 - if (llc->state == LLC_CONN_STATE_ADM) 50 + if (unlikely(llc->state == LLC_CONN_STATE_ADM)) 51 51 goto out; 52 52 rc = -EBUSY; 53 - if (llc_data_accept_state(llc->state)) { /* data_conn_refuse */ 54 - llc->failed_data_req = 1; 55 - goto out; 56 - } 57 - if (llc->p_flag) { 53 + if (unlikely(llc_data_accept_state(llc->state) || /* data_conn_refuse */ 54 + llc->p_flag)) { 58 55 llc->failed_data_req = 1; 59 56 goto out; 60 57 } ··· 107 110 ev->type = LLC_CONN_EV_TYPE_PRIM; 108 111 ev->prim = LLC_CONN_PRIM; 109 112 ev->prim_type = LLC_PRIM_TYPE_REQ; 113 + skb_set_owner_w(skb, sk); 110 114 rc = llc_conn_state_process(sk, skb); 111 115 } 112 116 out_put: ··· 142 144 skb = alloc_skb(0, GFP_ATOMIC); 143 145 if (!skb) 144 146 goto out; 147 + skb_set_owner_w(skb, sk); 145 148 sk->sk_state = TCP_CLOSING; 146 149 ev = llc_conn_ev(skb); 147 150 ev->type = LLC_CONN_EV_TYPE_PRIM;
+14 -5
net/llc/llc_input.c
··· 99 99 static inline int llc_fixup_skb(struct sk_buff *skb) 100 100 { 101 101 u8 llc_len = 2; 102 - struct llc_pdu_sn *pdu; 102 + struct llc_pdu_un *pdu; 103 103 104 - if (!pskb_may_pull(skb, sizeof(*pdu))) 104 + if (unlikely(!pskb_may_pull(skb, sizeof(*pdu)))) 105 105 return 0; 106 106 107 - pdu = (struct llc_pdu_sn *)skb->data; 107 + pdu = (struct llc_pdu_un *)skb->data; 108 108 if ((pdu->ctrl_1 & LLC_PDU_TYPE_MASK) == LLC_PDU_TYPE_U) 109 109 llc_len = 1; 110 110 llc_len += 2; 111 + 112 + if (unlikely(!pskb_may_pull(skb, llc_len))) 113 + return 0; 114 + 111 115 skb->h.raw += llc_len; 112 116 skb_pull(skb, llc_len); 113 117 if (skb->protocol == htons(ETH_P_802_2)) { ··· 170 166 */ 171 167 if (sap->rcv_func) { 172 168 sap->rcv_func(skb, dev, pt, orig_dev); 173 - goto out; 169 + goto out_put; 174 170 } 175 171 dest = llc_pdu_type(skb); 176 172 if (unlikely(!dest || !llc_type_handlers[dest - 1])) 177 - goto drop; 173 + goto drop_put; 178 174 llc_type_handlers[dest - 1](sap, skb); 175 + out_put: 176 + llc_sap_put(sap); 179 177 out: 180 178 return 0; 181 179 drop: 182 180 kfree_skb(skb); 183 181 goto out; 182 + drop_put: 183 + kfree_skb(skb); 184 + goto out_put; 184 185 handle_station: 185 186 if (!llc_station_handler) 186 187 goto drop;
+1 -1
net/llc/llc_output.c
··· 98 98 dsap, LLC_PDU_CMD); 99 99 llc_pdu_init_as_ui_cmd(skb); 100 100 rc = llc_mac_hdr_init(skb, skb->dev->dev_addr, dmac); 101 - if (!rc) 101 + if (likely(!rc)) 102 102 rc = dev_queue_xmit(skb); 103 103 return rc; 104 104 }
+1 -1
net/llc/llc_proc.c
··· 134 134 llc_ui_format_mac(seq, llc->daddr.mac); 135 135 seq_printf(seq, "@%02X %8d %8d %2d %3d %4d\n", llc->daddr.lsap, 136 136 atomic_read(&sk->sk_wmem_alloc), 137 - atomic_read(&sk->sk_rmem_alloc), 137 + atomic_read(&sk->sk_rmem_alloc) - llc->copied_seq, 138 138 sk->sk_state, 139 139 sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_uid : -1, 140 140 llc->link);
+7 -9
net/llc/llc_s_ac.c
··· 58 58 ev->daddr.lsap, LLC_PDU_CMD); 59 59 llc_pdu_init_as_ui_cmd(skb); 60 60 rc = llc_mac_hdr_init(skb, ev->saddr.mac, ev->daddr.mac); 61 - if (!rc) 61 + if (likely(!rc)) 62 62 rc = dev_queue_xmit(skb); 63 63 return rc; 64 64 } ··· 81 81 ev->daddr.lsap, LLC_PDU_CMD); 82 82 llc_pdu_init_as_xid_cmd(skb, LLC_XID_NULL_CLASS_2, 0); 83 83 rc = llc_mac_hdr_init(skb, ev->saddr.mac, ev->daddr.mac); 84 - if (!rc) 84 + if (likely(!rc)) 85 85 rc = dev_queue_xmit(skb); 86 86 return rc; 87 87 } ··· 103 103 llc_pdu_decode_sa(skb, mac_da); 104 104 llc_pdu_decode_da(skb, mac_sa); 105 105 llc_pdu_decode_ssap(skb, &dsap); 106 - nskb = llc_alloc_frame(); 106 + nskb = llc_alloc_frame(NULL, skb->dev); 107 107 if (!nskb) 108 108 goto out; 109 - nskb->dev = skb->dev; 110 109 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, dsap, 111 110 LLC_PDU_RSP); 112 111 llc_pdu_init_as_xid_rsp(nskb, LLC_XID_NULL_CLASS_2, 0); 113 112 rc = llc_mac_hdr_init(nskb, mac_sa, mac_da); 114 - if (!rc) 113 + if (likely(!rc)) 115 114 rc = dev_queue_xmit(nskb); 116 115 out: 117 116 return rc; ··· 134 135 ev->daddr.lsap, LLC_PDU_CMD); 135 136 llc_pdu_init_as_test_cmd(skb); 136 137 rc = llc_mac_hdr_init(skb, ev->saddr.mac, ev->daddr.mac); 137 - if (!rc) 138 + if (likely(!rc)) 138 139 rc = dev_queue_xmit(skb); 139 140 return rc; 140 141 } ··· 148 149 llc_pdu_decode_sa(skb, mac_da); 149 150 llc_pdu_decode_da(skb, mac_sa); 150 151 llc_pdu_decode_ssap(skb, &dsap); 151 - nskb = llc_alloc_frame(); 152 + nskb = llc_alloc_frame(NULL, skb->dev); 152 153 if (!nskb) 153 154 goto out; 154 - nskb->dev = skb->dev; 155 155 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, sap->laddr.lsap, dsap, 156 156 LLC_PDU_RSP); 157 157 llc_pdu_init_as_test_rsp(nskb, skb); 158 158 rc = llc_mac_hdr_init(nskb, mac_sa, mac_da); 159 - if (!rc) 159 + if (likely(!rc)) 160 160 rc = dev_queue_xmit(nskb); 161 161 out: 162 162 return rc;
+13 -7
net/llc/llc_sap.c
··· 26 26 27 27 /** 28 28 * llc_alloc_frame - allocates sk_buff for frame 29 + * @dev: network device this skb will be sent over 29 30 * 30 31 * Allocates an sk_buff for frame and initializes sk_buff fields. 31 32 * Returns allocated skb or %NULL when out of memory. 32 33 */ 33 - struct sk_buff *llc_alloc_frame(void) 34 + struct sk_buff *llc_alloc_frame(struct sock *sk, struct net_device *dev) 34 35 { 35 36 struct sk_buff *skb = alloc_skb(128, GFP_ATOMIC); 36 37 ··· 39 38 skb_reserve(skb, 50); 40 39 skb->nh.raw = skb->h.raw = skb->data; 41 40 skb->protocol = htons(ETH_P_802_2); 42 - skb->dev = dev_base->next; 41 + skb->dev = dev; 43 42 skb->mac.raw = skb->head; 43 + if (sk != NULL) 44 + skb_set_owner_w(skb, sk); 44 45 } 45 46 return skb; 46 47 } 47 48 48 - void llc_save_primitive(struct sk_buff* skb, u8 prim) 49 + void llc_save_primitive(struct sock *sk, struct sk_buff* skb, u8 prim) 49 50 { 50 - struct sockaddr_llc *addr = llc_ui_skb_cb(skb); 51 + struct sockaddr_llc *addr; 51 52 53 + if (skb->sk->sk_type == SOCK_STREAM) /* See UNIX98 */ 54 + return; 52 55 /* save primitive for use by the user. */ 53 - addr->sllc_family = skb->sk->sk_family; 56 + addr = llc_ui_skb_cb(skb); 57 + addr->sllc_family = sk->sk_family; 54 58 addr->sllc_arphrd = skb->dev->type; 55 59 addr->sllc_test = prim == LLC_TEST_PRIM; 56 60 addr->sllc_xid = prim == LLC_XID_PRIM; ··· 195 189 if (skb->sk->sk_state == TCP_LISTEN) 196 190 kfree_skb(skb); 197 191 else { 198 - llc_save_primitive(skb, ev->prim); 192 + llc_save_primitive(skb->sk, skb, ev->prim); 199 193 200 194 /* queue skb to the user. */ 201 195 if (sock_queue_rcv_skb(skb->sk, skb)) ··· 314 308 315 309 sk = llc_lookup_dgram(sap, &laddr); 316 310 if (sk) { 317 - skb->sk = sk; 311 + skb_set_owner_r(skb, sk); 318 312 llc_sap_rcv(sap, skb); 319 313 sock_put(sk); 320 314 } else
+14 -11
net/llc/llc_station.c
··· 50 50 struct sk_buff_head mac_pdu_q; 51 51 }; 52 52 53 + #define LLC_STATION_ACK_TIME (3 * HZ) 54 + 55 + int sysctl_llc_station_ack_timeout = LLC_STATION_ACK_TIME; 56 + 53 57 /* Types of events (possible values in 'ev->type') */ 54 58 #define LLC_STATION_EV_TYPE_SIMPLE 1 55 59 #define LLC_STATION_EV_TYPE_CONDITION 2 ··· 222 218 223 219 static int llc_station_ac_start_ack_timer(struct sk_buff *skb) 224 220 { 225 - mod_timer(&llc_main_station.ack_timer, jiffies + LLC_ACK_TIME * HZ); 221 + mod_timer(&llc_main_station.ack_timer, 222 + jiffies + sysctl_llc_station_ack_timeout); 226 223 return 0; 227 224 } 228 225 ··· 254 249 static int llc_station_ac_send_null_dsap_xid_c(struct sk_buff *skb) 255 250 { 256 251 int rc = 1; 257 - struct sk_buff *nskb = llc_alloc_frame(); 252 + struct sk_buff *nskb = llc_alloc_frame(NULL, skb->dev); 258 253 259 254 if (!nskb) 260 255 goto out; 261 256 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, 0, 0, LLC_PDU_CMD); 262 257 llc_pdu_init_as_xid_cmd(nskb, LLC_XID_NULL_CLASS_2, 127); 263 258 rc = llc_mac_hdr_init(nskb, llc_station_mac_sa, llc_station_mac_sa); 264 - if (rc) 259 + if (unlikely(rc)) 265 260 goto free; 266 261 llc_station_send_pdu(nskb); 267 262 out: ··· 275 270 { 276 271 u8 mac_da[ETH_ALEN], dsap; 277 272 int rc = 1; 278 - struct sk_buff* nskb = llc_alloc_frame(); 273 + struct sk_buff* nskb = llc_alloc_frame(NULL, skb->dev); 279 274 280 275 if (!nskb) 281 276 goto out; 282 277 rc = 0; 283 - nskb->dev = skb->dev; 284 278 llc_pdu_decode_sa(skb, mac_da); 285 279 llc_pdu_decode_ssap(skb, &dsap); 286 280 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, 0, dsap, LLC_PDU_RSP); 287 281 llc_pdu_init_as_xid_rsp(nskb, LLC_XID_NULL_CLASS_2, 127); 288 282 rc = llc_mac_hdr_init(nskb, llc_station_mac_sa, mac_da); 289 - if (rc) 283 + if (unlikely(rc)) 290 284 goto free; 291 285 llc_station_send_pdu(nskb); 292 286 out: ··· 299 295 { 300 296 u8 mac_da[ETH_ALEN], dsap; 301 297 int rc = 1; 302 - struct sk_buff *nskb = llc_alloc_frame(); 298 + struct sk_buff *nskb = llc_alloc_frame(NULL, skb->dev); 303 299 304 300 if (!nskb) 305 301 goto out; 306 302 rc = 0; 307 - nskb->dev = skb->dev; 308 303 llc_pdu_decode_sa(skb, mac_da); 309 304 llc_pdu_decode_ssap(skb, &dsap); 310 305 llc_pdu_header_init(nskb, LLC_PDU_TYPE_U, 0, dsap, LLC_PDU_RSP); 311 306 llc_pdu_init_as_test_rsp(nskb, skb); 312 307 rc = llc_mac_hdr_init(nskb, llc_station_mac_sa, mac_da); 313 - if (rc) 308 + if (unlikely(rc)) 314 309 goto free; 315 310 llc_station_send_pdu(nskb); 316 311 out: ··· 692 689 init_timer(&llc_main_station.ack_timer); 693 690 llc_main_station.ack_timer.data = (unsigned long)&llc_main_station; 694 691 llc_main_station.ack_timer.function = llc_station_ack_tmr_cb; 695 - 692 + llc_main_station.ack_timer.expires = jiffies + 693 + sysctl_llc_station_ack_timeout; 696 694 skb = alloc_skb(0, GFP_ATOMIC); 697 695 if (!skb) 698 696 goto out; ··· 701 697 llc_set_station_handler(llc_station_rcv); 702 698 ev = llc_station_ev(skb); 703 699 memset(ev, 0, sizeof(*ev)); 704 - llc_main_station.ack_timer.expires = jiffies + 3 * HZ; 705 700 llc_main_station.maximum_retry = 1; 706 701 llc_main_station.state = LLC_STATION_STATE_DOWN; 707 702 ev->type = LLC_STATION_EV_TYPE_SIMPLE;
+131
net/llc/sysctl_net_llc.c
··· 1 + /* 2 + * sysctl_net_llc.c: sysctl interface to LLC net subsystem. 3 + * 4 + * Arnaldo Carvalho de Melo <acme@conectiva.com.br> 5 + */ 6 + 7 + #include <linux/config.h> 8 + #include <linux/mm.h> 9 + #include <linux/init.h> 10 + #include <linux/sysctl.h> 11 + #include <net/llc.h> 12 + 13 + #ifndef CONFIG_SYSCTL 14 + #error This file should not be compiled without CONFIG_SYSCTL defined 15 + #endif 16 + 17 + static struct ctl_table llc2_timeout_table[] = { 18 + { 19 + .ctl_name = NET_LLC2_ACK_TIMEOUT, 20 + .procname = "ack", 21 + .data = &sysctl_llc2_ack_timeout, 22 + .maxlen = sizeof(long), 23 + .mode = 0644, 24 + .proc_handler = &proc_dointvec_jiffies, 25 + .strategy = &sysctl_jiffies, 26 + }, 27 + { 28 + .ctl_name = NET_LLC2_BUSY_TIMEOUT, 29 + .procname = "busy", 30 + .data = &sysctl_llc2_busy_timeout, 31 + .maxlen = sizeof(long), 32 + .mode = 0644, 33 + .proc_handler = &proc_dointvec_jiffies, 34 + .strategy = &sysctl_jiffies, 35 + }, 36 + { 37 + .ctl_name = NET_LLC2_P_TIMEOUT, 38 + .procname = "p", 39 + .data = &sysctl_llc2_p_timeout, 40 + .maxlen = sizeof(long), 41 + .mode = 0644, 42 + .proc_handler = &proc_dointvec_jiffies, 43 + .strategy = &sysctl_jiffies, 44 + }, 45 + { 46 + .ctl_name = NET_LLC2_REJ_TIMEOUT, 47 + .procname = "rej", 48 + .data = &sysctl_llc2_rej_timeout, 49 + .maxlen = sizeof(long), 50 + .mode = 0644, 51 + .proc_handler = &proc_dointvec_jiffies, 52 + .strategy = &sysctl_jiffies, 53 + }, 54 + { 0 }, 55 + }; 56 + 57 + static struct ctl_table llc_station_table[] = { 58 + { 59 + .ctl_name = NET_LLC_STATION_ACK_TIMEOUT, 60 + .procname = "ack_timeout", 61 + .data = &sysctl_llc_station_ack_timeout, 62 + .maxlen = sizeof(long), 63 + .mode = 0644, 64 + .proc_handler = &proc_dointvec_jiffies, 65 + .strategy = &sysctl_jiffies, 66 + }, 67 + { 0 }, 68 + }; 69 + 70 + static struct ctl_table llc2_dir_timeout_table[] = { 71 + { 72 + .ctl_name = NET_LLC2, 73 + .procname = "timeout", 74 + .mode = 0555, 75 + .child = llc2_timeout_table, 76 + }, 77 + { 0 }, 78 + }; 79 + 80 + static struct ctl_table llc_table[] = { 81 + { 82 + .ctl_name = NET_LLC2, 83 + .procname = "llc2", 84 + .mode = 0555, 85 + .child = llc2_dir_timeout_table, 86 + }, 87 + { 88 + .ctl_name = NET_LLC_STATION, 89 + .procname = "station", 90 + .mode = 0555, 91 + .child = llc_station_table, 92 + }, 93 + { 0 }, 94 + }; 95 + 96 + static struct ctl_table llc_dir_table[] = { 97 + { 98 + .ctl_name = NET_LLC, 99 + .procname = "llc", 100 + .mode = 0555, 101 + .child = llc_table, 102 + }, 103 + { 0 }, 104 + }; 105 + 106 + static struct ctl_table llc_root_table[] = { 107 + { 108 + .ctl_name = CTL_NET, 109 + .procname = "net", 110 + .mode = 0555, 111 + .child = llc_dir_table, 112 + }, 113 + { 0 }, 114 + }; 115 + 116 + static struct ctl_table_header *llc_table_header; 117 + 118 + int __init llc_sysctl_init(void) 119 + { 120 + llc_table_header = register_sysctl_table(llc_root_table, 1); 121 + 122 + return llc_table_header ? 0 : -ENOMEM; 123 + } 124 + 125 + void llc_sysctl_exit(void) 126 + { 127 + if (llc_table_header) { 128 + unregister_sysctl_table(llc_table_header); 129 + llc_table_header = NULL; 130 + } 131 + }
-6
net/packet/af_packet.c
··· 761 761 if (dev->hard_header) { 762 762 int res; 763 763 err = -EINVAL; 764 - if (saddr) { 765 - if (saddr->sll_halen != dev->addr_len) 766 - goto out_free; 767 - if (saddr->sll_hatype != dev->type) 768 - goto out_free; 769 - } 770 764 res = dev->hard_header(skb, dev, ntohs(proto), addr, NULL, len); 771 765 if (sock->type != SOCK_DGRAM) { 772 766 skb->tail = skb->data;
+3 -1
net/socket.c
··· 1700 1700 struct socket *sock; 1701 1701 char address[MAX_SOCK_ADDR]; 1702 1702 struct iovec iovstack[UIO_FASTIOV], *iov = iovstack; 1703 - unsigned char ctl[sizeof(struct cmsghdr) + 20]; /* 20 is size of ipv6_pktinfo */ 1703 + unsigned char ctl[sizeof(struct cmsghdr) + 20] 1704 + __attribute__ ((aligned (sizeof(__kernel_size_t)))); 1705 + /* 20 is size of ipv6_pktinfo */ 1704 1706 unsigned char *ctl_buf = ctl; 1705 1707 struct msghdr msg_sys; 1706 1708 int err, ctl_len, iov_size, total_len;