Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

[NETLINK]: af_netlink.c checkpatch cleanups

Fix large number of checkpatch errors.

Signed-off-by: Patrick McHardy <kaber@trash.net>
Signed-off-by: David S. Miller <davem@davemloft.net>

authored by

Patrick McHardy and committed by
David S. Miller
6ac552fd 2fcb45b6

+43 -38
+43 -38
net/netlink/af_netlink.c
··· 156 156 skb_queue_purge(&sk->sk_receive_queue); 157 157 158 158 if (!sock_flag(sk, SOCK_DEAD)) { 159 - printk("Freeing alive netlink socket %p\n", sk); 159 + printk(KERN_ERR "Freeing alive netlink socket %p\n", sk); 160 160 return; 161 161 } 162 162 BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc)); ··· 164 164 BUG_TRAP(!nlk_sk(sk)->groups); 165 165 } 166 166 167 - /* This lock without WQ_FLAG_EXCLUSIVE is good on UP and it is _very_ bad on SMP. 168 - * Look, when several writers sleep and reader wakes them up, all but one 167 + /* This lock without WQ_FLAG_EXCLUSIVE is good on UP and it is _very_ bad on 168 + * SMP. Look, when several writers sleep and reader wakes them up, all but one 169 169 * immediately hit write lock and grab all the cpus. Exclusive sleep solves 170 170 * this, _but_ remember, it adds useless work on UP machines. 171 171 */ ··· 178 178 DECLARE_WAITQUEUE(wait, current); 179 179 180 180 add_wait_queue_exclusive(&nl_table_wait, &wait); 181 - for(;;) { 181 + for (;;) { 182 182 set_current_state(TASK_UNINTERRUPTIBLE); 183 183 if (atomic_read(&nl_table_users) == 0) 184 184 break; ··· 192 192 } 193 193 } 194 194 195 - static __inline__ void netlink_table_ungrab(void) 195 + static inline void netlink_table_ungrab(void) 196 196 { 197 197 write_unlock_irq(&nl_table_lock); 198 198 wake_up(&nl_table_wait); 199 199 } 200 200 201 - static __inline__ void 201 + static inline void 202 202 netlink_lock_table(void) 203 203 { 204 204 /* read_lock() synchronizes us to netlink_table_grab */ ··· 208 208 read_unlock(&nl_table_lock); 209 209 } 210 210 211 - static __inline__ void 211 + static inline void 212 212 netlink_unlock_table(void) 213 213 { 214 214 if (atomic_dec_and_test(&nl_table_users)) 215 215 wake_up(&nl_table_wait); 216 216 } 217 217 218 - static __inline__ struct sock *netlink_lookup(struct net *net, int protocol, u32 pid) 218 + static inline struct sock *netlink_lookup(struct net *net, int protocol, 219 + u32 pid) 219 220 { 220 221 struct nl_pid_hash *hash = &nl_table[protocol].hash; 221 222 struct hlist_head *head; ··· 429 428 if (sock->type != SOCK_RAW && sock->type != SOCK_DGRAM) 430 429 return -ESOCKTNOSUPPORT; 431 430 432 - if (protocol<0 || protocol >= MAX_LINKS) 431 + if (protocol < 0 || protocol >= MAX_LINKS) 433 432 return -EPROTONOSUPPORT; 434 433 435 434 netlink_lock_table(); ··· 446 445 cb_mutex = nl_table[protocol].cb_mutex; 447 446 netlink_unlock_table(); 448 447 449 - if ((err = __netlink_create(net, sock, cb_mutex, protocol)) < 0) 448 + err = __netlink_create(net, sock, cb_mutex, protocol); 449 + if (err < 0) 450 450 goto out_module; 451 451 452 452 nlk = nlk_sk(sock->sk); ··· 592 590 err = -ENOMEM; 593 591 goto out_unlock; 594 592 } 595 - memset((char*)new_groups + NLGRPSZ(nlk->ngroups), 0, 593 + memset((char *)new_groups + NLGRPSZ(nlk->ngroups), 0, 596 594 NLGRPSZ(groups) - NLGRPSZ(nlk->ngroups)); 597 595 598 596 nlk->groups = new_groups; ··· 602 600 return err; 603 601 } 604 602 605 - static int netlink_bind(struct socket *sock, struct sockaddr *addr, int addr_len) 603 + static int netlink_bind(struct socket *sock, struct sockaddr *addr, 604 + int addr_len) 606 605 { 607 606 struct sock *sk = sock->sk; 608 607 struct net *net = sk->sk_net; ··· 654 651 int err = 0; 655 652 struct sock *sk = sock->sk; 656 653 struct netlink_sock *nlk = nlk_sk(sk); 657 - struct sockaddr_nl *nladdr=(struct sockaddr_nl*)addr; 654 + struct sockaddr_nl *nladdr = (struct sockaddr_nl *)addr; 658 655 659 656 if (addr->sa_family == AF_UNSPEC) { 660 657 sk->sk_state = NETLINK_UNCONNECTED; ··· 681 678 return err; 682 679 } 683 680 684 - static int netlink_getname(struct socket *sock, struct sockaddr *addr, int *addr_len, int peer) 681 + static int netlink_getname(struct socket *sock, struct sockaddr *addr, 682 + int *addr_len, int peer) 685 683 { 686 684 struct sock *sk = sock->sk; 687 685 struct netlink_sock *nlk = nlk_sk(sk); 688 - struct sockaddr_nl *nladdr=(struct sockaddr_nl *)addr; 686 + struct sockaddr_nl *nladdr = (struct sockaddr_nl *)addr; 689 687 690 688 nladdr->nl_family = AF_NETLINK; 691 689 nladdr->nl_pad = 0; ··· 889 885 890 886 return netlink_sendskb(sk, skb); 891 887 } 888 + EXPORT_SYMBOL(netlink_unicast); 892 889 893 890 int netlink_has_listeners(struct sock *sk, unsigned int group) 894 891 { ··· 910 905 } 911 906 EXPORT_SYMBOL_GPL(netlink_has_listeners); 912 907 913 - static __inline__ int netlink_broadcast_deliver(struct sock *sk, struct sk_buff *skb) 908 + static inline int netlink_broadcast_deliver(struct sock *sk, 909 + struct sk_buff *skb) 914 910 { 915 911 struct netlink_sock *nlk = nlk_sk(sk); 916 912 ··· 1032 1026 return -ENOBUFS; 1033 1027 return -ESRCH; 1034 1028 } 1029 + EXPORT_SYMBOL(netlink_broadcast); 1035 1030 1036 1031 struct netlink_set_err_data { 1037 1032 struct sock *exclude_sk; ··· 1189 1182 struct sock_iocb *siocb = kiocb_to_siocb(kiocb); 1190 1183 struct sock *sk = sock->sk; 1191 1184 struct netlink_sock *nlk = nlk_sk(sk); 1192 - struct sockaddr_nl *addr=msg->msg_name; 1185 + struct sockaddr_nl *addr = msg->msg_name; 1193 1186 u32 dst_pid; 1194 1187 u32 dst_group; 1195 1188 struct sk_buff *skb; ··· 1228 1221 goto out; 1229 1222 err = -ENOBUFS; 1230 1223 skb = alloc_skb(len, GFP_KERNEL); 1231 - if (skb==NULL) 1224 + if (skb == NULL) 1232 1225 goto out; 1233 1226 1234 1227 NETLINK_CB(skb).pid = nlk->pid; ··· 1244 1237 */ 1245 1238 1246 1239 err = -EFAULT; 1247 - if (memcpy_fromiovec(skb_put(skb,len), msg->msg_iov, len)) { 1240 + if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)) { 1248 1241 kfree_skb(skb); 1249 1242 goto out; 1250 1243 } ··· 1283 1276 1284 1277 copied = 0; 1285 1278 1286 - skb = skb_recv_datagram(sk,flags,noblock,&err); 1287 - if (skb==NULL) 1279 + skb = skb_recv_datagram(sk, flags, noblock, &err); 1280 + if (skb == NULL) 1288 1281 goto out; 1289 1282 1290 1283 msg->msg_namelen = 0; ··· 1299 1292 err = skb_copy_datagram_iovec(skb, 0, msg->msg_iov, copied); 1300 1293 1301 1294 if (msg->msg_name) { 1302 - struct sockaddr_nl *addr = (struct sockaddr_nl*)msg->msg_name; 1295 + struct sockaddr_nl *addr = (struct sockaddr_nl *)msg->msg_name; 1303 1296 addr->nl_family = AF_NETLINK; 1304 1297 addr->nl_pad = 0; 1305 1298 addr->nl_pid = NETLINK_CB(skb).pid; ··· 1351 1344 1352 1345 BUG_ON(!nl_table); 1353 1346 1354 - if (unit<0 || unit>=MAX_LINKS) 1347 + if (unit < 0 || unit >= MAX_LINKS) 1355 1348 return NULL; 1356 1349 1357 1350 if (sock_create_lite(PF_NETLINK, SOCK_DGRAM, unit, &sock)) ··· 1397 1390 sock_release(sock); 1398 1391 return NULL; 1399 1392 } 1393 + EXPORT_SYMBOL(netlink_kernel_create); 1400 1394 1401 1395 /** 1402 1396 * netlink_change_ngroups - change number of multicast groups ··· 1469 1461 if ((unsigned int)protocol < MAX_LINKS) 1470 1462 nl_table[protocol].nl_nonroot = flags; 1471 1463 } 1464 + EXPORT_SYMBOL(netlink_set_nonroot); 1472 1465 1473 1466 static void netlink_destroy_callback(struct netlink_callback *cb) 1474 1467 { ··· 1538 1529 1539 1530 int netlink_dump_start(struct sock *ssk, struct sk_buff *skb, 1540 1531 struct nlmsghdr *nlh, 1541 - int (*dump)(struct sk_buff *skb, struct netlink_callback*), 1542 - int (*done)(struct netlink_callback*)) 1532 + int (*dump)(struct sk_buff *skb, 1533 + struct netlink_callback *), 1534 + int (*done)(struct netlink_callback *)) 1543 1535 { 1544 1536 struct netlink_callback *cb; 1545 1537 struct sock *sk; ··· 1581 1571 */ 1582 1572 return -EINTR; 1583 1573 } 1574 + EXPORT_SYMBOL(netlink_dump_start); 1584 1575 1585 1576 void netlink_ack(struct sk_buff *in_skb, struct nlmsghdr *nlh, int err) 1586 1577 { ··· 1616 1605 memcpy(&errmsg->msg, nlh, err ? nlh->nlmsg_len : sizeof(*nlh)); 1617 1606 netlink_unicast(in_skb->sk, skb, NETLINK_CB(in_skb).pid, MSG_DONTWAIT); 1618 1607 } 1608 + EXPORT_SYMBOL(netlink_ack); 1619 1609 1620 1610 int netlink_rcv_skb(struct sk_buff *skb, int (*cb)(struct sk_buff *, 1621 1611 struct nlmsghdr *)) ··· 1650 1638 netlink_ack(skb, nlh, err); 1651 1639 1652 1640 skip: 1653 - msglen = NLMSG_ALIGN(nlh->nlmsg_len); 1641 + msglen = NLMSG_ALIGN(nlh->nlmsg_len); 1654 1642 if (msglen > skb->len) 1655 1643 msglen = skb->len; 1656 1644 skb_pull(skb, msglen); ··· 1658 1646 1659 1647 return 0; 1660 1648 } 1649 + EXPORT_SYMBOL(netlink_rcv_skb); 1661 1650 1662 1651 /** 1663 1652 * nlmsg_notify - send a notification netlink message ··· 1691 1678 1692 1679 return err; 1693 1680 } 1681 + EXPORT_SYMBOL(nlmsg_notify); 1694 1682 1695 1683 #ifdef CONFIG_PROC_FS 1696 1684 struct nl_seq_iter { ··· 1708 1694 struct hlist_node *node; 1709 1695 loff_t off = 0; 1710 1696 1711 - for (i=0; i<MAX_LINKS; i++) { 1697 + for (i = 0; i < MAX_LINKS; i++) { 1712 1698 struct nl_pid_hash *hash = &nl_table[i].hash; 1713 1699 1714 1700 for (j = 0; j <= hash->mask; j++) { ··· 1834 1820 { 1835 1821 return atomic_notifier_chain_register(&netlink_chain, nb); 1836 1822 } 1823 + EXPORT_SYMBOL(netlink_register_notifier); 1837 1824 1838 1825 int netlink_unregister_notifier(struct notifier_block *nb) 1839 1826 { 1840 1827 return atomic_notifier_chain_unregister(&netlink_chain, nb); 1841 1828 } 1829 + EXPORT_SYMBOL(netlink_unregister_notifier); 1842 1830 1843 1831 static const struct proto_ops netlink_ops = { 1844 1832 .family = PF_NETLINK, ··· 1945 1929 } 1946 1930 1947 1931 core_initcall(netlink_proto_init); 1948 - 1949 - EXPORT_SYMBOL(netlink_ack); 1950 - EXPORT_SYMBOL(netlink_rcv_skb); 1951 - EXPORT_SYMBOL(netlink_broadcast); 1952 - EXPORT_SYMBOL(netlink_dump_start); 1953 - EXPORT_SYMBOL(netlink_kernel_create); 1954 - EXPORT_SYMBOL(netlink_register_notifier); 1955 - EXPORT_SYMBOL(netlink_set_nonroot); 1956 - EXPORT_SYMBOL(netlink_unicast); 1957 - EXPORT_SYMBOL(netlink_unregister_notifier); 1958 - EXPORT_SYMBOL(nlmsg_notify);