Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v2.6.21-rc4 1198 lines 31 kB view raw
1/* 2 * IPVS An implementation of the IP virtual server support for the 3 * LINUX operating system. IPVS is now implemented as a module 4 * over the Netfilter framework. IPVS can be used to build a 5 * high-performance and highly available server based on a 6 * cluster of servers. 7 * 8 * Version: $Id: ip_vs_core.c,v 1.34 2003/05/10 03:05:23 wensong Exp $ 9 * 10 * Authors: Wensong Zhang <wensong@linuxvirtualserver.org> 11 * Peter Kese <peter.kese@ijs.si> 12 * Julian Anastasov <ja@ssi.bg> 13 * 14 * This program is free software; you can redistribute it and/or 15 * modify it under the terms of the GNU General Public License 16 * as published by the Free Software Foundation; either version 17 * 2 of the License, or (at your option) any later version. 18 * 19 * The IPVS code for kernel 2.2 was done by Wensong Zhang and Peter Kese, 20 * with changes/fixes from Julian Anastasov, Lars Marowsky-Bree, Horms 21 * and others. 22 * 23 * Changes: 24 * Paul `Rusty' Russell properly handle non-linear skbs 25 * Harald Welte don't use nfcache 26 * 27 */ 28 29#include <linux/module.h> 30#include <linux/kernel.h> 31#include <linux/ip.h> 32#include <linux/tcp.h> 33#include <linux/icmp.h> 34 35#include <net/ip.h> 36#include <net/tcp.h> 37#include <net/udp.h> 38#include <net/icmp.h> /* for icmp_send */ 39#include <net/route.h> 40 41#include <linux/netfilter.h> 42#include <linux/netfilter_ipv4.h> 43 44#include <net/ip_vs.h> 45 46 47EXPORT_SYMBOL(register_ip_vs_scheduler); 48EXPORT_SYMBOL(unregister_ip_vs_scheduler); 49EXPORT_SYMBOL(ip_vs_skb_replace); 50EXPORT_SYMBOL(ip_vs_proto_name); 51EXPORT_SYMBOL(ip_vs_conn_new); 52EXPORT_SYMBOL(ip_vs_conn_in_get); 53EXPORT_SYMBOL(ip_vs_conn_out_get); 54#ifdef CONFIG_IP_VS_PROTO_TCP 55EXPORT_SYMBOL(ip_vs_tcp_conn_listen); 56#endif 57EXPORT_SYMBOL(ip_vs_conn_put); 58#ifdef CONFIG_IP_VS_DEBUG 59EXPORT_SYMBOL(ip_vs_get_debug_level); 60#endif 61EXPORT_SYMBOL(ip_vs_make_skb_writable); 62 63 64/* ID used in ICMP lookups */ 65#define icmp_id(icmph) (((icmph)->un).echo.id) 66 67const char *ip_vs_proto_name(unsigned proto) 68{ 69 static char buf[20]; 70 71 switch (proto) { 72 case IPPROTO_IP: 73 return "IP"; 74 case IPPROTO_UDP: 75 return "UDP"; 76 case IPPROTO_TCP: 77 return "TCP"; 78 case IPPROTO_ICMP: 79 return "ICMP"; 80 default: 81 sprintf(buf, "IP_%d", proto); 82 return buf; 83 } 84} 85 86void ip_vs_init_hash_table(struct list_head *table, int rows) 87{ 88 while (--rows >= 0) 89 INIT_LIST_HEAD(&table[rows]); 90} 91 92static inline void 93ip_vs_in_stats(struct ip_vs_conn *cp, struct sk_buff *skb) 94{ 95 struct ip_vs_dest *dest = cp->dest; 96 if (dest && (dest->flags & IP_VS_DEST_F_AVAILABLE)) { 97 spin_lock(&dest->stats.lock); 98 dest->stats.inpkts++; 99 dest->stats.inbytes += skb->len; 100 spin_unlock(&dest->stats.lock); 101 102 spin_lock(&dest->svc->stats.lock); 103 dest->svc->stats.inpkts++; 104 dest->svc->stats.inbytes += skb->len; 105 spin_unlock(&dest->svc->stats.lock); 106 107 spin_lock(&ip_vs_stats.lock); 108 ip_vs_stats.inpkts++; 109 ip_vs_stats.inbytes += skb->len; 110 spin_unlock(&ip_vs_stats.lock); 111 } 112} 113 114 115static inline void 116ip_vs_out_stats(struct ip_vs_conn *cp, struct sk_buff *skb) 117{ 118 struct ip_vs_dest *dest = cp->dest; 119 if (dest && (dest->flags & IP_VS_DEST_F_AVAILABLE)) { 120 spin_lock(&dest->stats.lock); 121 dest->stats.outpkts++; 122 dest->stats.outbytes += skb->len; 123 spin_unlock(&dest->stats.lock); 124 125 spin_lock(&dest->svc->stats.lock); 126 dest->svc->stats.outpkts++; 127 dest->svc->stats.outbytes += skb->len; 128 spin_unlock(&dest->svc->stats.lock); 129 130 spin_lock(&ip_vs_stats.lock); 131 ip_vs_stats.outpkts++; 132 ip_vs_stats.outbytes += skb->len; 133 spin_unlock(&ip_vs_stats.lock); 134 } 135} 136 137 138static inline void 139ip_vs_conn_stats(struct ip_vs_conn *cp, struct ip_vs_service *svc) 140{ 141 spin_lock(&cp->dest->stats.lock); 142 cp->dest->stats.conns++; 143 spin_unlock(&cp->dest->stats.lock); 144 145 spin_lock(&svc->stats.lock); 146 svc->stats.conns++; 147 spin_unlock(&svc->stats.lock); 148 149 spin_lock(&ip_vs_stats.lock); 150 ip_vs_stats.conns++; 151 spin_unlock(&ip_vs_stats.lock); 152} 153 154 155static inline int 156ip_vs_set_state(struct ip_vs_conn *cp, int direction, 157 const struct sk_buff *skb, 158 struct ip_vs_protocol *pp) 159{ 160 if (unlikely(!pp->state_transition)) 161 return 0; 162 return pp->state_transition(cp, direction, skb, pp); 163} 164 165 166int ip_vs_make_skb_writable(struct sk_buff **pskb, int writable_len) 167{ 168 struct sk_buff *skb = *pskb; 169 170 /* skb is already used, better copy skb and its payload */ 171 if (unlikely(skb_shared(skb) || skb->sk)) 172 goto copy_skb; 173 174 /* skb data is already used, copy it */ 175 if (unlikely(skb_cloned(skb))) 176 goto copy_data; 177 178 return pskb_may_pull(skb, writable_len); 179 180 copy_data: 181 if (unlikely(writable_len > skb->len)) 182 return 0; 183 return !pskb_expand_head(skb, 0, 0, GFP_ATOMIC); 184 185 copy_skb: 186 if (unlikely(writable_len > skb->len)) 187 return 0; 188 skb = skb_copy(skb, GFP_ATOMIC); 189 if (!skb) 190 return 0; 191 BUG_ON(skb_is_nonlinear(skb)); 192 193 /* Rest of kernel will get very unhappy if we pass it a 194 suddenly-orphaned skbuff */ 195 if ((*pskb)->sk) 196 skb_set_owner_w(skb, (*pskb)->sk); 197 kfree_skb(*pskb); 198 *pskb = skb; 199 return 1; 200} 201 202/* 203 * IPVS persistent scheduling function 204 * It creates a connection entry according to its template if exists, 205 * or selects a server and creates a connection entry plus a template. 206 * Locking: we are svc user (svc->refcnt), so we hold all dests too 207 * Protocols supported: TCP, UDP 208 */ 209static struct ip_vs_conn * 210ip_vs_sched_persist(struct ip_vs_service *svc, 211 const struct sk_buff *skb, 212 __be16 ports[2]) 213{ 214 struct ip_vs_conn *cp = NULL; 215 struct iphdr *iph = skb->nh.iph; 216 struct ip_vs_dest *dest; 217 struct ip_vs_conn *ct; 218 __be16 dport; /* destination port to forward */ 219 __be32 snet; /* source network of the client, after masking */ 220 221 /* Mask saddr with the netmask to adjust template granularity */ 222 snet = iph->saddr & svc->netmask; 223 224 IP_VS_DBG(6, "p-schedule: src %u.%u.%u.%u:%u dest %u.%u.%u.%u:%u " 225 "mnet %u.%u.%u.%u\n", 226 NIPQUAD(iph->saddr), ntohs(ports[0]), 227 NIPQUAD(iph->daddr), ntohs(ports[1]), 228 NIPQUAD(snet)); 229 230 /* 231 * As far as we know, FTP is a very complicated network protocol, and 232 * it uses control connection and data connections. For active FTP, 233 * FTP server initialize data connection to the client, its source port 234 * is often 20. For passive FTP, FTP server tells the clients the port 235 * that it passively listens to, and the client issues the data 236 * connection. In the tunneling or direct routing mode, the load 237 * balancer is on the client-to-server half of connection, the port 238 * number is unknown to the load balancer. So, a conn template like 239 * <caddr, 0, vaddr, 0, daddr, 0> is created for persistent FTP 240 * service, and a template like <caddr, 0, vaddr, vport, daddr, dport> 241 * is created for other persistent services. 242 */ 243 if (ports[1] == svc->port) { 244 /* Check if a template already exists */ 245 if (svc->port != FTPPORT) 246 ct = ip_vs_ct_in_get(iph->protocol, snet, 0, 247 iph->daddr, ports[1]); 248 else 249 ct = ip_vs_ct_in_get(iph->protocol, snet, 0, 250 iph->daddr, 0); 251 252 if (!ct || !ip_vs_check_template(ct)) { 253 /* 254 * No template found or the dest of the connection 255 * template is not available. 256 */ 257 dest = svc->scheduler->schedule(svc, skb); 258 if (dest == NULL) { 259 IP_VS_DBG(1, "p-schedule: no dest found.\n"); 260 return NULL; 261 } 262 263 /* 264 * Create a template like <protocol,caddr,0, 265 * vaddr,vport,daddr,dport> for non-ftp service, 266 * and <protocol,caddr,0,vaddr,0,daddr,0> 267 * for ftp service. 268 */ 269 if (svc->port != FTPPORT) 270 ct = ip_vs_conn_new(iph->protocol, 271 snet, 0, 272 iph->daddr, 273 ports[1], 274 dest->addr, dest->port, 275 IP_VS_CONN_F_TEMPLATE, 276 dest); 277 else 278 ct = ip_vs_conn_new(iph->protocol, 279 snet, 0, 280 iph->daddr, 0, 281 dest->addr, 0, 282 IP_VS_CONN_F_TEMPLATE, 283 dest); 284 if (ct == NULL) 285 return NULL; 286 287 ct->timeout = svc->timeout; 288 } else { 289 /* set destination with the found template */ 290 dest = ct->dest; 291 } 292 dport = dest->port; 293 } else { 294 /* 295 * Note: persistent fwmark-based services and persistent 296 * port zero service are handled here. 297 * fwmark template: <IPPROTO_IP,caddr,0,fwmark,0,daddr,0> 298 * port zero template: <protocol,caddr,0,vaddr,0,daddr,0> 299 */ 300 if (svc->fwmark) 301 ct = ip_vs_ct_in_get(IPPROTO_IP, snet, 0, 302 htonl(svc->fwmark), 0); 303 else 304 ct = ip_vs_ct_in_get(iph->protocol, snet, 0, 305 iph->daddr, 0); 306 307 if (!ct || !ip_vs_check_template(ct)) { 308 /* 309 * If it is not persistent port zero, return NULL, 310 * otherwise create a connection template. 311 */ 312 if (svc->port) 313 return NULL; 314 315 dest = svc->scheduler->schedule(svc, skb); 316 if (dest == NULL) { 317 IP_VS_DBG(1, "p-schedule: no dest found.\n"); 318 return NULL; 319 } 320 321 /* 322 * Create a template according to the service 323 */ 324 if (svc->fwmark) 325 ct = ip_vs_conn_new(IPPROTO_IP, 326 snet, 0, 327 htonl(svc->fwmark), 0, 328 dest->addr, 0, 329 IP_VS_CONN_F_TEMPLATE, 330 dest); 331 else 332 ct = ip_vs_conn_new(iph->protocol, 333 snet, 0, 334 iph->daddr, 0, 335 dest->addr, 0, 336 IP_VS_CONN_F_TEMPLATE, 337 dest); 338 if (ct == NULL) 339 return NULL; 340 341 ct->timeout = svc->timeout; 342 } else { 343 /* set destination with the found template */ 344 dest = ct->dest; 345 } 346 dport = ports[1]; 347 } 348 349 /* 350 * Create a new connection according to the template 351 */ 352 cp = ip_vs_conn_new(iph->protocol, 353 iph->saddr, ports[0], 354 iph->daddr, ports[1], 355 dest->addr, dport, 356 0, 357 dest); 358 if (cp == NULL) { 359 ip_vs_conn_put(ct); 360 return NULL; 361 } 362 363 /* 364 * Add its control 365 */ 366 ip_vs_control_add(cp, ct); 367 ip_vs_conn_put(ct); 368 369 ip_vs_conn_stats(cp, svc); 370 return cp; 371} 372 373 374/* 375 * IPVS main scheduling function 376 * It selects a server according to the virtual service, and 377 * creates a connection entry. 378 * Protocols supported: TCP, UDP 379 */ 380struct ip_vs_conn * 381ip_vs_schedule(struct ip_vs_service *svc, const struct sk_buff *skb) 382{ 383 struct ip_vs_conn *cp = NULL; 384 struct iphdr *iph = skb->nh.iph; 385 struct ip_vs_dest *dest; 386 __be16 _ports[2], *pptr; 387 388 pptr = skb_header_pointer(skb, iph->ihl*4, 389 sizeof(_ports), _ports); 390 if (pptr == NULL) 391 return NULL; 392 393 /* 394 * Persistent service 395 */ 396 if (svc->flags & IP_VS_SVC_F_PERSISTENT) 397 return ip_vs_sched_persist(svc, skb, pptr); 398 399 /* 400 * Non-persistent service 401 */ 402 if (!svc->fwmark && pptr[1] != svc->port) { 403 if (!svc->port) 404 IP_VS_ERR("Schedule: port zero only supported " 405 "in persistent services, " 406 "check your ipvs configuration\n"); 407 return NULL; 408 } 409 410 dest = svc->scheduler->schedule(svc, skb); 411 if (dest == NULL) { 412 IP_VS_DBG(1, "Schedule: no dest found.\n"); 413 return NULL; 414 } 415 416 /* 417 * Create a connection entry. 418 */ 419 cp = ip_vs_conn_new(iph->protocol, 420 iph->saddr, pptr[0], 421 iph->daddr, pptr[1], 422 dest->addr, dest->port?dest->port:pptr[1], 423 0, 424 dest); 425 if (cp == NULL) 426 return NULL; 427 428 IP_VS_DBG(6, "Schedule fwd:%c c:%u.%u.%u.%u:%u v:%u.%u.%u.%u:%u " 429 "d:%u.%u.%u.%u:%u conn->flags:%X conn->refcnt:%d\n", 430 ip_vs_fwd_tag(cp), 431 NIPQUAD(cp->caddr), ntohs(cp->cport), 432 NIPQUAD(cp->vaddr), ntohs(cp->vport), 433 NIPQUAD(cp->daddr), ntohs(cp->dport), 434 cp->flags, atomic_read(&cp->refcnt)); 435 436 ip_vs_conn_stats(cp, svc); 437 return cp; 438} 439 440 441/* 442 * Pass or drop the packet. 443 * Called by ip_vs_in, when the virtual service is available but 444 * no destination is available for a new connection. 445 */ 446int ip_vs_leave(struct ip_vs_service *svc, struct sk_buff *skb, 447 struct ip_vs_protocol *pp) 448{ 449 __be16 _ports[2], *pptr; 450 struct iphdr *iph = skb->nh.iph; 451 452 pptr = skb_header_pointer(skb, iph->ihl*4, 453 sizeof(_ports), _ports); 454 if (pptr == NULL) { 455 ip_vs_service_put(svc); 456 return NF_DROP; 457 } 458 459 /* if it is fwmark-based service, the cache_bypass sysctl is up 460 and the destination is RTN_UNICAST (and not local), then create 461 a cache_bypass connection entry */ 462 if (sysctl_ip_vs_cache_bypass && svc->fwmark 463 && (inet_addr_type(iph->daddr) == RTN_UNICAST)) { 464 int ret, cs; 465 struct ip_vs_conn *cp; 466 467 ip_vs_service_put(svc); 468 469 /* create a new connection entry */ 470 IP_VS_DBG(6, "ip_vs_leave: create a cache_bypass entry\n"); 471 cp = ip_vs_conn_new(iph->protocol, 472 iph->saddr, pptr[0], 473 iph->daddr, pptr[1], 474 0, 0, 475 IP_VS_CONN_F_BYPASS, 476 NULL); 477 if (cp == NULL) 478 return NF_DROP; 479 480 /* statistics */ 481 ip_vs_in_stats(cp, skb); 482 483 /* set state */ 484 cs = ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pp); 485 486 /* transmit the first SYN packet */ 487 ret = cp->packet_xmit(skb, cp, pp); 488 /* do not touch skb anymore */ 489 490 atomic_inc(&cp->in_pkts); 491 ip_vs_conn_put(cp); 492 return ret; 493 } 494 495 /* 496 * When the virtual ftp service is presented, packets destined 497 * for other services on the VIP may get here (except services 498 * listed in the ipvs table), pass the packets, because it is 499 * not ipvs job to decide to drop the packets. 500 */ 501 if ((svc->port == FTPPORT) && (pptr[1] != FTPPORT)) { 502 ip_vs_service_put(svc); 503 return NF_ACCEPT; 504 } 505 506 ip_vs_service_put(svc); 507 508 /* 509 * Notify the client that the destination is unreachable, and 510 * release the socket buffer. 511 * Since it is in IP layer, the TCP socket is not actually 512 * created, the TCP RST packet cannot be sent, instead that 513 * ICMP_PORT_UNREACH is sent here no matter it is TCP/UDP. --WZ 514 */ 515 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_PORT_UNREACH, 0); 516 return NF_DROP; 517} 518 519 520/* 521 * It is hooked before NF_IP_PRI_NAT_SRC at the NF_IP_POST_ROUTING 522 * chain, and is used for VS/NAT. 523 * It detects packets for VS/NAT connections and sends the packets 524 * immediately. This can avoid that iptable_nat mangles the packets 525 * for VS/NAT. 526 */ 527static unsigned int ip_vs_post_routing(unsigned int hooknum, 528 struct sk_buff **pskb, 529 const struct net_device *in, 530 const struct net_device *out, 531 int (*okfn)(struct sk_buff *)) 532{ 533 if (!((*pskb)->ipvs_property)) 534 return NF_ACCEPT; 535 /* The packet was sent from IPVS, exit this chain */ 536 return NF_STOP; 537} 538 539__sum16 ip_vs_checksum_complete(struct sk_buff *skb, int offset) 540{ 541 return csum_fold(skb_checksum(skb, offset, skb->len - offset, 0)); 542} 543 544static inline struct sk_buff * 545ip_vs_gather_frags(struct sk_buff *skb, u_int32_t user) 546{ 547 skb = ip_defrag(skb, user); 548 if (skb) 549 ip_send_check(skb->nh.iph); 550 return skb; 551} 552 553/* 554 * Packet has been made sufficiently writable in caller 555 * - inout: 1=in->out, 0=out->in 556 */ 557void ip_vs_nat_icmp(struct sk_buff *skb, struct ip_vs_protocol *pp, 558 struct ip_vs_conn *cp, int inout) 559{ 560 struct iphdr *iph = skb->nh.iph; 561 unsigned int icmp_offset = iph->ihl*4; 562 struct icmphdr *icmph = (struct icmphdr *)(skb->nh.raw + icmp_offset); 563 struct iphdr *ciph = (struct iphdr *)(icmph + 1); 564 565 if (inout) { 566 iph->saddr = cp->vaddr; 567 ip_send_check(iph); 568 ciph->daddr = cp->vaddr; 569 ip_send_check(ciph); 570 } else { 571 iph->daddr = cp->daddr; 572 ip_send_check(iph); 573 ciph->saddr = cp->daddr; 574 ip_send_check(ciph); 575 } 576 577 /* the TCP/UDP port */ 578 if (IPPROTO_TCP == ciph->protocol || IPPROTO_UDP == ciph->protocol) { 579 __be16 *ports = (void *)ciph + ciph->ihl*4; 580 581 if (inout) 582 ports[1] = cp->vport; 583 else 584 ports[0] = cp->dport; 585 } 586 587 /* And finally the ICMP checksum */ 588 icmph->checksum = 0; 589 icmph->checksum = ip_vs_checksum_complete(skb, icmp_offset); 590 skb->ip_summed = CHECKSUM_UNNECESSARY; 591 592 if (inout) 593 IP_VS_DBG_PKT(11, pp, skb, (void *)ciph - (void *)iph, 594 "Forwarding altered outgoing ICMP"); 595 else 596 IP_VS_DBG_PKT(11, pp, skb, (void *)ciph - (void *)iph, 597 "Forwarding altered incoming ICMP"); 598} 599 600/* 601 * Handle ICMP messages in the inside-to-outside direction (outgoing). 602 * Find any that might be relevant, check against existing connections, 603 * forward to the right destination host if relevant. 604 * Currently handles error types - unreachable, quench, ttl exceeded. 605 * (Only used in VS/NAT) 606 */ 607static int ip_vs_out_icmp(struct sk_buff **pskb, int *related) 608{ 609 struct sk_buff *skb = *pskb; 610 struct iphdr *iph; 611 struct icmphdr _icmph, *ic; 612 struct iphdr _ciph, *cih; /* The ip header contained within the ICMP */ 613 struct ip_vs_conn *cp; 614 struct ip_vs_protocol *pp; 615 unsigned int offset, ihl, verdict; 616 617 *related = 1; 618 619 /* reassemble IP fragments */ 620 if (skb->nh.iph->frag_off & __constant_htons(IP_MF|IP_OFFSET)) { 621 skb = ip_vs_gather_frags(skb, IP_DEFRAG_VS_OUT); 622 if (!skb) 623 return NF_STOLEN; 624 *pskb = skb; 625 } 626 627 iph = skb->nh.iph; 628 offset = ihl = iph->ihl * 4; 629 ic = skb_header_pointer(skb, offset, sizeof(_icmph), &_icmph); 630 if (ic == NULL) 631 return NF_DROP; 632 633 IP_VS_DBG(12, "Outgoing ICMP (%d,%d) %u.%u.%u.%u->%u.%u.%u.%u\n", 634 ic->type, ntohs(icmp_id(ic)), 635 NIPQUAD(iph->saddr), NIPQUAD(iph->daddr)); 636 637 /* 638 * Work through seeing if this is for us. 639 * These checks are supposed to be in an order that means easy 640 * things are checked first to speed up processing.... however 641 * this means that some packets will manage to get a long way 642 * down this stack and then be rejected, but that's life. 643 */ 644 if ((ic->type != ICMP_DEST_UNREACH) && 645 (ic->type != ICMP_SOURCE_QUENCH) && 646 (ic->type != ICMP_TIME_EXCEEDED)) { 647 *related = 0; 648 return NF_ACCEPT; 649 } 650 651 /* Now find the contained IP header */ 652 offset += sizeof(_icmph); 653 cih = skb_header_pointer(skb, offset, sizeof(_ciph), &_ciph); 654 if (cih == NULL) 655 return NF_ACCEPT; /* The packet looks wrong, ignore */ 656 657 pp = ip_vs_proto_get(cih->protocol); 658 if (!pp) 659 return NF_ACCEPT; 660 661 /* Is the embedded protocol header present? */ 662 if (unlikely(cih->frag_off & __constant_htons(IP_OFFSET) && 663 pp->dont_defrag)) 664 return NF_ACCEPT; 665 666 IP_VS_DBG_PKT(11, pp, skb, offset, "Checking outgoing ICMP for"); 667 668 offset += cih->ihl * 4; 669 670 /* The embedded headers contain source and dest in reverse order */ 671 cp = pp->conn_out_get(skb, pp, cih, offset, 1); 672 if (!cp) 673 return NF_ACCEPT; 674 675 verdict = NF_DROP; 676 677 if (IP_VS_FWD_METHOD(cp) != 0) { 678 IP_VS_ERR("shouldn't reach here, because the box is on the" 679 "half connection in the tun/dr module.\n"); 680 } 681 682 /* Ensure the checksum is correct */ 683 if (skb->ip_summed != CHECKSUM_UNNECESSARY && 684 ip_vs_checksum_complete(skb, ihl)) { 685 /* Failed checksum! */ 686 IP_VS_DBG(1, "Forward ICMP: failed checksum from %d.%d.%d.%d!\n", 687 NIPQUAD(iph->saddr)); 688 goto out; 689 } 690 691 if (IPPROTO_TCP == cih->protocol || IPPROTO_UDP == cih->protocol) 692 offset += 2 * sizeof(__u16); 693 if (!ip_vs_make_skb_writable(pskb, offset)) 694 goto out; 695 skb = *pskb; 696 697 ip_vs_nat_icmp(skb, pp, cp, 1); 698 699 /* do the statistics and put it back */ 700 ip_vs_out_stats(cp, skb); 701 702 skb->ipvs_property = 1; 703 verdict = NF_ACCEPT; 704 705 out: 706 __ip_vs_conn_put(cp); 707 708 return verdict; 709} 710 711static inline int is_tcp_reset(const struct sk_buff *skb) 712{ 713 struct tcphdr _tcph, *th; 714 715 th = skb_header_pointer(skb, skb->nh.iph->ihl * 4, 716 sizeof(_tcph), &_tcph); 717 if (th == NULL) 718 return 0; 719 return th->rst; 720} 721 722/* 723 * It is hooked at the NF_IP_FORWARD chain, used only for VS/NAT. 724 * Check if outgoing packet belongs to the established ip_vs_conn, 725 * rewrite addresses of the packet and send it on its way... 726 */ 727static unsigned int 728ip_vs_out(unsigned int hooknum, struct sk_buff **pskb, 729 const struct net_device *in, const struct net_device *out, 730 int (*okfn)(struct sk_buff *)) 731{ 732 struct sk_buff *skb = *pskb; 733 struct iphdr *iph; 734 struct ip_vs_protocol *pp; 735 struct ip_vs_conn *cp; 736 int ihl; 737 738 EnterFunction(11); 739 740 if (skb->ipvs_property) 741 return NF_ACCEPT; 742 743 iph = skb->nh.iph; 744 if (unlikely(iph->protocol == IPPROTO_ICMP)) { 745 int related, verdict = ip_vs_out_icmp(pskb, &related); 746 747 if (related) 748 return verdict; 749 skb = *pskb; 750 iph = skb->nh.iph; 751 } 752 753 pp = ip_vs_proto_get(iph->protocol); 754 if (unlikely(!pp)) 755 return NF_ACCEPT; 756 757 /* reassemble IP fragments */ 758 if (unlikely(iph->frag_off & __constant_htons(IP_MF|IP_OFFSET) && 759 !pp->dont_defrag)) { 760 skb = ip_vs_gather_frags(skb, IP_DEFRAG_VS_OUT); 761 if (!skb) 762 return NF_STOLEN; 763 iph = skb->nh.iph; 764 *pskb = skb; 765 } 766 767 ihl = iph->ihl << 2; 768 769 /* 770 * Check if the packet belongs to an existing entry 771 */ 772 cp = pp->conn_out_get(skb, pp, iph, ihl, 0); 773 774 if (unlikely(!cp)) { 775 if (sysctl_ip_vs_nat_icmp_send && 776 (pp->protocol == IPPROTO_TCP || 777 pp->protocol == IPPROTO_UDP)) { 778 __be16 _ports[2], *pptr; 779 780 pptr = skb_header_pointer(skb, ihl, 781 sizeof(_ports), _ports); 782 if (pptr == NULL) 783 return NF_ACCEPT; /* Not for me */ 784 if (ip_vs_lookup_real_service(iph->protocol, 785 iph->saddr, pptr[0])) { 786 /* 787 * Notify the real server: there is no 788 * existing entry if it is not RST 789 * packet or not TCP packet. 790 */ 791 if (iph->protocol != IPPROTO_TCP 792 || !is_tcp_reset(skb)) { 793 icmp_send(skb,ICMP_DEST_UNREACH, 794 ICMP_PORT_UNREACH, 0); 795 return NF_DROP; 796 } 797 } 798 } 799 IP_VS_DBG_PKT(12, pp, skb, 0, 800 "packet continues traversal as normal"); 801 return NF_ACCEPT; 802 } 803 804 IP_VS_DBG_PKT(11, pp, skb, 0, "Outgoing packet"); 805 806 if (!ip_vs_make_skb_writable(pskb, ihl)) 807 goto drop; 808 809 /* mangle the packet */ 810 if (pp->snat_handler && !pp->snat_handler(pskb, pp, cp)) 811 goto drop; 812 skb = *pskb; 813 skb->nh.iph->saddr = cp->vaddr; 814 ip_send_check(skb->nh.iph); 815 816 /* For policy routing, packets originating from this 817 * machine itself may be routed differently to packets 818 * passing through. We want this packet to be routed as 819 * if it came from this machine itself. So re-compute 820 * the routing information. 821 */ 822 if (ip_route_me_harder(pskb, RTN_LOCAL) != 0) 823 goto drop; 824 skb = *pskb; 825 826 IP_VS_DBG_PKT(10, pp, skb, 0, "After SNAT"); 827 828 ip_vs_out_stats(cp, skb); 829 ip_vs_set_state(cp, IP_VS_DIR_OUTPUT, skb, pp); 830 ip_vs_conn_put(cp); 831 832 skb->ipvs_property = 1; 833 834 LeaveFunction(11); 835 return NF_ACCEPT; 836 837 drop: 838 ip_vs_conn_put(cp); 839 kfree_skb(*pskb); 840 return NF_STOLEN; 841} 842 843 844/* 845 * Handle ICMP messages in the outside-to-inside direction (incoming). 846 * Find any that might be relevant, check against existing connections, 847 * forward to the right destination host if relevant. 848 * Currently handles error types - unreachable, quench, ttl exceeded. 849 */ 850static int 851ip_vs_in_icmp(struct sk_buff **pskb, int *related, unsigned int hooknum) 852{ 853 struct sk_buff *skb = *pskb; 854 struct iphdr *iph; 855 struct icmphdr _icmph, *ic; 856 struct iphdr _ciph, *cih; /* The ip header contained within the ICMP */ 857 struct ip_vs_conn *cp; 858 struct ip_vs_protocol *pp; 859 unsigned int offset, ihl, verdict; 860 861 *related = 1; 862 863 /* reassemble IP fragments */ 864 if (skb->nh.iph->frag_off & __constant_htons(IP_MF|IP_OFFSET)) { 865 skb = ip_vs_gather_frags(skb, 866 hooknum == NF_IP_LOCAL_IN ? 867 IP_DEFRAG_VS_IN : IP_DEFRAG_VS_FWD); 868 if (!skb) 869 return NF_STOLEN; 870 *pskb = skb; 871 } 872 873 iph = skb->nh.iph; 874 offset = ihl = iph->ihl * 4; 875 ic = skb_header_pointer(skb, offset, sizeof(_icmph), &_icmph); 876 if (ic == NULL) 877 return NF_DROP; 878 879 IP_VS_DBG(12, "Incoming ICMP (%d,%d) %u.%u.%u.%u->%u.%u.%u.%u\n", 880 ic->type, ntohs(icmp_id(ic)), 881 NIPQUAD(iph->saddr), NIPQUAD(iph->daddr)); 882 883 /* 884 * Work through seeing if this is for us. 885 * These checks are supposed to be in an order that means easy 886 * things are checked first to speed up processing.... however 887 * this means that some packets will manage to get a long way 888 * down this stack and then be rejected, but that's life. 889 */ 890 if ((ic->type != ICMP_DEST_UNREACH) && 891 (ic->type != ICMP_SOURCE_QUENCH) && 892 (ic->type != ICMP_TIME_EXCEEDED)) { 893 *related = 0; 894 return NF_ACCEPT; 895 } 896 897 /* Now find the contained IP header */ 898 offset += sizeof(_icmph); 899 cih = skb_header_pointer(skb, offset, sizeof(_ciph), &_ciph); 900 if (cih == NULL) 901 return NF_ACCEPT; /* The packet looks wrong, ignore */ 902 903 pp = ip_vs_proto_get(cih->protocol); 904 if (!pp) 905 return NF_ACCEPT; 906 907 /* Is the embedded protocol header present? */ 908 if (unlikely(cih->frag_off & __constant_htons(IP_OFFSET) && 909 pp->dont_defrag)) 910 return NF_ACCEPT; 911 912 IP_VS_DBG_PKT(11, pp, skb, offset, "Checking incoming ICMP for"); 913 914 offset += cih->ihl * 4; 915 916 /* The embedded headers contain source and dest in reverse order */ 917 cp = pp->conn_in_get(skb, pp, cih, offset, 1); 918 if (!cp) 919 return NF_ACCEPT; 920 921 verdict = NF_DROP; 922 923 /* Ensure the checksum is correct */ 924 if (skb->ip_summed != CHECKSUM_UNNECESSARY && 925 ip_vs_checksum_complete(skb, ihl)) { 926 /* Failed checksum! */ 927 IP_VS_DBG(1, "Incoming ICMP: failed checksum from %d.%d.%d.%d!\n", 928 NIPQUAD(iph->saddr)); 929 goto out; 930 } 931 932 /* do the statistics and put it back */ 933 ip_vs_in_stats(cp, skb); 934 if (IPPROTO_TCP == cih->protocol || IPPROTO_UDP == cih->protocol) 935 offset += 2 * sizeof(__u16); 936 verdict = ip_vs_icmp_xmit(skb, cp, pp, offset); 937 /* do not touch skb anymore */ 938 939 out: 940 __ip_vs_conn_put(cp); 941 942 return verdict; 943} 944 945/* 946 * Check if it's for virtual services, look it up, 947 * and send it on its way... 948 */ 949static unsigned int 950ip_vs_in(unsigned int hooknum, struct sk_buff **pskb, 951 const struct net_device *in, const struct net_device *out, 952 int (*okfn)(struct sk_buff *)) 953{ 954 struct sk_buff *skb = *pskb; 955 struct iphdr *iph; 956 struct ip_vs_protocol *pp; 957 struct ip_vs_conn *cp; 958 int ret, restart; 959 int ihl; 960 961 /* 962 * Big tappo: only PACKET_HOST (neither loopback nor mcasts) 963 * ... don't know why 1st test DOES NOT include 2nd (?) 964 */ 965 if (unlikely(skb->pkt_type != PACKET_HOST 966 || skb->dev == &loopback_dev || skb->sk)) { 967 IP_VS_DBG(12, "packet type=%d proto=%d daddr=%d.%d.%d.%d ignored\n", 968 skb->pkt_type, 969 skb->nh.iph->protocol, 970 NIPQUAD(skb->nh.iph->daddr)); 971 return NF_ACCEPT; 972 } 973 974 iph = skb->nh.iph; 975 if (unlikely(iph->protocol == IPPROTO_ICMP)) { 976 int related, verdict = ip_vs_in_icmp(pskb, &related, hooknum); 977 978 if (related) 979 return verdict; 980 skb = *pskb; 981 iph = skb->nh.iph; 982 } 983 984 /* Protocol supported? */ 985 pp = ip_vs_proto_get(iph->protocol); 986 if (unlikely(!pp)) 987 return NF_ACCEPT; 988 989 ihl = iph->ihl << 2; 990 991 /* 992 * Check if the packet belongs to an existing connection entry 993 */ 994 cp = pp->conn_in_get(skb, pp, iph, ihl, 0); 995 996 if (unlikely(!cp)) { 997 int v; 998 999 if (!pp->conn_schedule(skb, pp, &v, &cp)) 1000 return v; 1001 } 1002 1003 if (unlikely(!cp)) { 1004 /* sorry, all this trouble for a no-hit :) */ 1005 IP_VS_DBG_PKT(12, pp, skb, 0, 1006 "packet continues traversal as normal"); 1007 return NF_ACCEPT; 1008 } 1009 1010 IP_VS_DBG_PKT(11, pp, skb, 0, "Incoming packet"); 1011 1012 /* Check the server status */ 1013 if (cp->dest && !(cp->dest->flags & IP_VS_DEST_F_AVAILABLE)) { 1014 /* the destination server is not available */ 1015 1016 if (sysctl_ip_vs_expire_nodest_conn) { 1017 /* try to expire the connection immediately */ 1018 ip_vs_conn_expire_now(cp); 1019 } 1020 /* don't restart its timer, and silently 1021 drop the packet. */ 1022 __ip_vs_conn_put(cp); 1023 return NF_DROP; 1024 } 1025 1026 ip_vs_in_stats(cp, skb); 1027 restart = ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pp); 1028 if (cp->packet_xmit) 1029 ret = cp->packet_xmit(skb, cp, pp); 1030 /* do not touch skb anymore */ 1031 else { 1032 IP_VS_DBG_RL("warning: packet_xmit is null"); 1033 ret = NF_ACCEPT; 1034 } 1035 1036 /* increase its packet counter and check if it is needed 1037 to be synchronized */ 1038 atomic_inc(&cp->in_pkts); 1039 if ((ip_vs_sync_state & IP_VS_STATE_MASTER) && 1040 (cp->protocol != IPPROTO_TCP || 1041 cp->state == IP_VS_TCP_S_ESTABLISHED) && 1042 (atomic_read(&cp->in_pkts) % sysctl_ip_vs_sync_threshold[1] 1043 == sysctl_ip_vs_sync_threshold[0])) 1044 ip_vs_sync_conn(cp); 1045 1046 ip_vs_conn_put(cp); 1047 return ret; 1048} 1049 1050 1051/* 1052 * It is hooked at the NF_IP_FORWARD chain, in order to catch ICMP 1053 * related packets destined for 0.0.0.0/0. 1054 * When fwmark-based virtual service is used, such as transparent 1055 * cache cluster, TCP packets can be marked and routed to ip_vs_in, 1056 * but ICMP destined for 0.0.0.0/0 cannot not be easily marked and 1057 * sent to ip_vs_in_icmp. So, catch them at the NF_IP_FORWARD chain 1058 * and send them to ip_vs_in_icmp. 1059 */ 1060static unsigned int 1061ip_vs_forward_icmp(unsigned int hooknum, struct sk_buff **pskb, 1062 const struct net_device *in, const struct net_device *out, 1063 int (*okfn)(struct sk_buff *)) 1064{ 1065 int r; 1066 1067 if ((*pskb)->nh.iph->protocol != IPPROTO_ICMP) 1068 return NF_ACCEPT; 1069 1070 return ip_vs_in_icmp(pskb, &r, hooknum); 1071} 1072 1073 1074/* After packet filtering, forward packet through VS/DR, VS/TUN, 1075 or VS/NAT(change destination), so that filtering rules can be 1076 applied to IPVS. */ 1077static struct nf_hook_ops ip_vs_in_ops = { 1078 .hook = ip_vs_in, 1079 .owner = THIS_MODULE, 1080 .pf = PF_INET, 1081 .hooknum = NF_IP_LOCAL_IN, 1082 .priority = 100, 1083}; 1084 1085/* After packet filtering, change source only for VS/NAT */ 1086static struct nf_hook_ops ip_vs_out_ops = { 1087 .hook = ip_vs_out, 1088 .owner = THIS_MODULE, 1089 .pf = PF_INET, 1090 .hooknum = NF_IP_FORWARD, 1091 .priority = 100, 1092}; 1093 1094/* After packet filtering (but before ip_vs_out_icmp), catch icmp 1095 destined for 0.0.0.0/0, which is for incoming IPVS connections */ 1096static struct nf_hook_ops ip_vs_forward_icmp_ops = { 1097 .hook = ip_vs_forward_icmp, 1098 .owner = THIS_MODULE, 1099 .pf = PF_INET, 1100 .hooknum = NF_IP_FORWARD, 1101 .priority = 99, 1102}; 1103 1104/* Before the netfilter connection tracking, exit from POST_ROUTING */ 1105static struct nf_hook_ops ip_vs_post_routing_ops = { 1106 .hook = ip_vs_post_routing, 1107 .owner = THIS_MODULE, 1108 .pf = PF_INET, 1109 .hooknum = NF_IP_POST_ROUTING, 1110 .priority = NF_IP_PRI_NAT_SRC-1, 1111}; 1112 1113 1114/* 1115 * Initialize IP Virtual Server 1116 */ 1117static int __init ip_vs_init(void) 1118{ 1119 int ret; 1120 1121 ret = ip_vs_control_init(); 1122 if (ret < 0) { 1123 IP_VS_ERR("can't setup control.\n"); 1124 goto cleanup_nothing; 1125 } 1126 1127 ip_vs_protocol_init(); 1128 1129 ret = ip_vs_app_init(); 1130 if (ret < 0) { 1131 IP_VS_ERR("can't setup application helper.\n"); 1132 goto cleanup_protocol; 1133 } 1134 1135 ret = ip_vs_conn_init(); 1136 if (ret < 0) { 1137 IP_VS_ERR("can't setup connection table.\n"); 1138 goto cleanup_app; 1139 } 1140 1141 ret = nf_register_hook(&ip_vs_in_ops); 1142 if (ret < 0) { 1143 IP_VS_ERR("can't register in hook.\n"); 1144 goto cleanup_conn; 1145 } 1146 1147 ret = nf_register_hook(&ip_vs_out_ops); 1148 if (ret < 0) { 1149 IP_VS_ERR("can't register out hook.\n"); 1150 goto cleanup_inops; 1151 } 1152 ret = nf_register_hook(&ip_vs_post_routing_ops); 1153 if (ret < 0) { 1154 IP_VS_ERR("can't register post_routing hook.\n"); 1155 goto cleanup_outops; 1156 } 1157 ret = nf_register_hook(&ip_vs_forward_icmp_ops); 1158 if (ret < 0) { 1159 IP_VS_ERR("can't register forward_icmp hook.\n"); 1160 goto cleanup_postroutingops; 1161 } 1162 1163 IP_VS_INFO("ipvs loaded.\n"); 1164 return ret; 1165 1166 cleanup_postroutingops: 1167 nf_unregister_hook(&ip_vs_post_routing_ops); 1168 cleanup_outops: 1169 nf_unregister_hook(&ip_vs_out_ops); 1170 cleanup_inops: 1171 nf_unregister_hook(&ip_vs_in_ops); 1172 cleanup_conn: 1173 ip_vs_conn_cleanup(); 1174 cleanup_app: 1175 ip_vs_app_cleanup(); 1176 cleanup_protocol: 1177 ip_vs_protocol_cleanup(); 1178 ip_vs_control_cleanup(); 1179 cleanup_nothing: 1180 return ret; 1181} 1182 1183static void __exit ip_vs_cleanup(void) 1184{ 1185 nf_unregister_hook(&ip_vs_forward_icmp_ops); 1186 nf_unregister_hook(&ip_vs_post_routing_ops); 1187 nf_unregister_hook(&ip_vs_out_ops); 1188 nf_unregister_hook(&ip_vs_in_ops); 1189 ip_vs_conn_cleanup(); 1190 ip_vs_app_cleanup(); 1191 ip_vs_protocol_cleanup(); 1192 ip_vs_control_cleanup(); 1193 IP_VS_INFO("ipvs unloaded.\n"); 1194} 1195 1196module_init(ip_vs_init); 1197module_exit(ip_vs_cleanup); 1198MODULE_LICENSE("GPL");