Willy Tarreau | d8e42b6 | 2015-08-18 21:51:36 +0200 | [diff] [blame] | 1 | ----------------------- |
| 2 | HAProxy Starter Guide |
| 3 | ----------------------- |
| 4 | version 1.6 |
| 5 | |
| 6 | |
| 7 | This document is an introduction to HAProxy for all those who don't know it, as |
| 8 | well as for those who want to re-discover it when they know older versions. Its |
| 9 | primary focus is to provide users with all the elements to decide if HAProxy is |
| 10 | the product they're looking for or not. Advanced users may find here some parts |
| 11 | of solutions to some ideas they had just because they were not aware of a given |
| 12 | new feature. Some sizing information are also provided, the product's lifecycle |
| 13 | is explained, and comparisons with partially overlapping products are provided. |
| 14 | |
| 15 | This document doesn't provide any configuration help nor hint, but it explains |
| 16 | where to find the relevant documents. The summary below is meant to help you |
| 17 | search sections by name and navigate through the document. |
| 18 | |
| 19 | Note to documentation contributors : |
| 20 | This document is formatted with 80 columns per line, with even number of |
| 21 | spaces for indentation and without tabs. Please follow these rules strictly |
| 22 | so that it remains easily printable everywhere. If you add sections, please |
| 23 | update the summary below for easier searching. |
| 24 | |
| 25 | |
| 26 | Summary |
| 27 | ------- |
| 28 | |
| 29 | 1. Available documentation |
| 30 | |
| 31 | 2. Quick introduction to load balancing and load balancers |
| 32 | |
| 33 | 3. Introduction to HAProxy |
| 34 | 3.1. What HAProxy is and is not |
| 35 | 3.2. How HAProxy works |
| 36 | 3.3. Basic features |
| 37 | 3.3.1. Proxying |
| 38 | 3.3.2. SSL |
| 39 | 3.3.3. Monitoring |
| 40 | 3.3.4. High availability |
| 41 | 3.3.5. Load balancing |
| 42 | 3.3.6. Stickiness |
| 43 | 3.3.7. Sampling and converting information |
| 44 | 3.3.8. Maps |
| 45 | 3.3.9. ACLs and conditions |
| 46 | 3.3.10. Content switching |
| 47 | 3.3.11. Stick-tables |
| 48 | 3.3.12. Formated strings |
| 49 | 3.3.13. HTTP rewriting and redirection |
| 50 | 3.3.14. Server protection |
| 51 | 3.3.15. Logging |
| 52 | 3.3.16. Statistics |
| 53 | 3.4. Advanced features |
| 54 | 3.4.1. Management |
| 55 | 3.4.2. System-specific capabilities |
| 56 | 3.4.3. Scripting |
| 57 | 3.5. Sizing |
| 58 | 3.6. How to get HAProxy |
| 59 | |
| 60 | 4. Companion products and alternatives |
| 61 | 4.1. Apache HTTP server |
| 62 | 4.2. NGINX |
| 63 | 4.3. Varnish |
| 64 | 4.4. Alternatives |
| 65 | |
| 66 | |
| 67 | 1. Available documentation |
| 68 | -------------------------- |
| 69 | |
| 70 | The complete HAProxy documentation is contained in the following documents. |
| 71 | Please ensure to consult the relevant documentation to save time and to get the |
| 72 | most accurate response to your needs. Also please refrain from sending questions |
| 73 | to the mailing list whose responses are present in these documents. |
| 74 | |
| 75 | - intro.txt (this document) : it presents the basics of load balancing, |
| 76 | HAProxy as a product, what it does, what it doesn't do, some known traps to |
| 77 | avoid, some OS-specific limitations, how to get it, how it evolves, how to |
| 78 | ensure you're running with all known fixes how to update it, complements and |
| 79 | alternatives. |
| 80 | |
| 81 | - management.txt : it explains how to start haproxy, how to manage it at |
| 82 | runtime, how to manage it on multiple nodes, how to proceed with seamless |
| 83 | upgrades. |
| 84 | |
| 85 | - configuration.txt : the reference manual details all configuration keywords |
| 86 | and their options. It is used when a configuration change is needed. |
| 87 | |
| 88 | - architecture.txt : the architecture manual explains how to best architect a |
| 89 | load-balanced infrastructure and how to interact with third party products. |
| 90 | |
| 91 | - coding-style.txt : this is for developers who want to propose some code to |
| 92 | the project. It explains the style to adopt for the code. It's not very |
| 93 | strict and not all the code base completely respects it but contributions |
| 94 | which diverge too much from it will be rejected. |
| 95 | |
| 96 | - proxy-protocol.txt : this is the de-facto specification of the PROXY |
| 97 | protocol which is implemented by HAProxy and a number of third party |
| 98 | products. |
| 99 | |
| 100 | - README : how to build haproxy from sources |
| 101 | |
| 102 | |
| 103 | 2. Quick introduction to load balancing and load balancers |
| 104 | ---------------------------------------------------------- |
| 105 | |
| 106 | Load balancing consists in aggregating multiple components in order to achieve |
| 107 | a total processing capacity above each component's individual capacity, without |
| 108 | any intervention from the end user and in a scalable way. This results in more |
| 109 | operations being performed simultaneously by the time it takes component to |
| 110 | perform only one. A single operation however will still be performed on a single |
| 111 | component at a time and will not get faster than without load balancing. It |
| 112 | always requires at least as many operations as available components and an |
| 113 | efficient load balancing mechanism to make use of all components and to fully |
| 114 | benefit from the load balancing. A good example of this is the number of lanes |
| 115 | on a highway which allows as many cars to pass during the same time frame |
| 116 | without increasing their individual speed. |
| 117 | |
| 118 | Examples of load balancing : |
| 119 | |
| 120 | - Process scheduling in multi-processor systems |
| 121 | - Link load balancing (eg: EtherChannel, Bonding) |
| 122 | - IP address load balancing (eg: ECMP, DNS roundrobin) |
| 123 | - Server load balancing (via load balancers) |
| 124 | |
| 125 | The mechanism or component which performs the load balancing operation is |
| 126 | called a load balancer. In web environments these components are called a |
| 127 | "network load balancer", and more commonly a "load balancer" given that this |
| 128 | activity is by far the best known case of load balancing. |
| 129 | |
| 130 | A load balancer may act : |
| 131 | |
| 132 | - at the link level : this is called link load balancing, and it consists in |
| 133 | chosing what network link to send a packet to; |
| 134 | |
| 135 | - at the network level : this is called network load balancing, and it |
| 136 | consists in chosing what route a series of packets will follow; |
| 137 | |
| 138 | - at the server level : this is called server load balancing and it consists |
| 139 | in deciding what server will process a connection or request. |
| 140 | |
| 141 | Two distinct technologies exist and address different needs, though with some |
| 142 | overlapping. |
| 143 | |
| 144 | The first one acts at the packet level and processes packets more or less |
| 145 | individually. There is a 1-to-1 relation between input and output packets, so |
| 146 | it is possible to follow the traffic on both sides of the load balancer using a |
| 147 | regular network sniffer. This technology can be very cheap and extremely fast. |
| 148 | It is usually implemented in hardware (ASICs) allowing to reach line rate, such |
| 149 | as switches doing ECMP. Usually stateless, it can also be stateful (consider |
| 150 | the session a packet belongs to and called layer4-LB or L4), may support DSR |
| 151 | (direct server return, without passing through the LB again) if the packets |
| 152 | were not modified, but provides almost no content awareness. This technology is |
| 153 | very well suited to network-level load balancing, though it is sometimes used |
| 154 | for very basic server load balancing at high speed. |
| 155 | |
| 156 | The second one acts on session contents. It requires that the input streams is |
| 157 | reassembled and processed as a whole. The contents may be modified, and the |
| 158 | output stream is segmented into new packets. For this reason it is generally |
| 159 | performed by proxies and they're often called layer 7 load balancers or L7. |
| 160 | This implies that there are two distinct connections on each side, and that |
| 161 | there is no relation between input and output packets sizes nor counts. Clients |
| 162 | and servers are not required to use the same protocol (for example IPv4 vs |
| 163 | IPv6, clear vs SSL). The operations are always stateful, and the return traffic |
| 164 | must pass through the load balancer. The extra processing comes with a cost so |
| 165 | it's not always possible to achieve line rate, especially with small packets. |
| 166 | On the other hand, it offers wide possibilities and is generally achieved by |
| 167 | pure software, even if embedded into hardware appliances. This technology is |
| 168 | very well suited for server load balancing. |
| 169 | |
| 170 | Packet-based load balancers are generally deployed in cut-through mode, so they |
| 171 | are installed on the normal path of the traffic and divert it according to the |
| 172 | configuration. The return traffic doesn't necessarily pass through the load |
| 173 | balancer. Some modifications may be applied to the network destination address |
| 174 | in order to direct the traffic to the proper destination. In this case, it is |
| 175 | mandatory that the return traffic passes through the load balancer. If the |
| 176 | routes doesn't make this possible, the load balancer may also replace the |
| 177 | packets' source address with its own in order to force the return traffic to |
| 178 | pass through it. |
| 179 | |
| 180 | Proxy-based load balancers are deployed as a server with their own IP address |
| 181 | and ports, without architecture changes. Sometimes this requires to perform some |
| 182 | adaptations to the applications so that clients are properly directed to the |
| 183 | load balancer's IP address and not directly to the server's. Some load balancers |
| 184 | may have to adjust some servers' responses to make this possible (eg: the HTTP |
| 185 | Location header field used in HTTP redirects). Some proxy-based load balancers |
| 186 | may intercept traffic for an address they don't own, and spoof the client's |
| 187 | address when connecting to the server. This allows them to be deployed as if |
| 188 | they were a regular router or firewall, in a cut-through mode very similar to |
| 189 | the packet based load balancers. This is particularly appreciated for products |
| 190 | which combine both packet mode and proxy mode. In this case DSR is obviously |
| 191 | still not possible and the return traffic still has to be routed back to the |
| 192 | load balancer. |
| 193 | |
| 194 | A very scalable layered approach would consist in having a front router which |
| 195 | receives traffic from multiple load balanced links, and uses ECMP to distribute |
| 196 | this traffic to a first layer of multiple stateful packet-based load balancers |
| 197 | (L4). These L4 load balancers in turn pass the traffic to an even larger number |
| 198 | of proxy-based load balancers (L7), which have to parse the contents to decide |
| 199 | what server will ultimately receive the traffic. |
| 200 | |
| 201 | The number of components and possible paths for the traffic increases the risk |
| 202 | of failure; in very large environments, it is even normal to permanently have |
| 203 | a few faulty components being fixed or replaced. Load balancing done without |
| 204 | awareness of the whole stack's health significantly degrades availability. For |
| 205 | this reason, any sane load balancer will verify that the components it intends |
| 206 | to deliver the traffic to are still alive and reachable, and it will stop |
| 207 | delivering traffic to faulty ones. This can be achieved using various methods. |
| 208 | |
| 209 | The most common one consists in periodically sending probes to ensure the |
| 210 | component is still operational. These probes are called "health checks". They |
| 211 | must be representative of the type of failure to address. For example a ping- |
| 212 | based check will not detect that a web server has crashed and doesn't listen to |
| 213 | a port anymore, while a connection to the port will verify this, and a more |
| 214 | advanced request may even validate that the server still works and that the |
| 215 | database it relies on is still accessible. Health checks often involve a few |
| 216 | retries to cover for occasional measuring errors. The period between checks |
| 217 | must be small enough to ensure the faulty component is not used for too long |
| 218 | after an error occurs. |
| 219 | |
| 220 | Other methods consist in sampling the production traffic sent to a destination |
| 221 | to observe if it is processed correctly or not, and to evince the components |
| 222 | which return inappropriate responses. However this requires to sacrify a part |
| 223 | of the production traffic and this is not always acceptable. A combination of |
| 224 | these two mechanisms provides the best of both worlds, with both of them being |
| 225 | used to detect a fault, and only health checks to detect the end of the fault. |
| 226 | A last method involves centralized reporting : a central monitoring agent |
| 227 | periodically updates all load balancers about all components' state. This gives |
| 228 | a global view of the infrastructure to all components, though sometimes with |
| 229 | less accuracy or responsiveness. It's best suited for environments with many |
| 230 | load balancers and many servers. |
| 231 | |
| 232 | Layer 7 load balancers also face another challenge known as stickiness or |
| 233 | persistence. The principle is that they generally have to direct multiple |
| 234 | subsequent requests or connections from a same origin (such as an end user) to |
| 235 | the same target. The best known example is the shopping cart on an online |
| 236 | store. If each click leads to a new connection, the user must always be sent |
| 237 | to the server which holds his shopping cart. Content-awareness makes it easier |
| 238 | to spot some elements in the request to identify the server to deliver it to, |
| 239 | but that's not always enough. For example if the source address is used as a |
| 240 | key to pick a server, it can be decided that a hash-based algorithm will be |
| 241 | used and that a given IP address will always be sent to the same server based |
| 242 | on a divide of the address by the number of available servers. But if one |
| 243 | server fails, the result changes and all users are suddenly sent to a different |
| 244 | server and lose their shopping cart. The solution against this issue consists |
| 245 | in memorizing the chosen target so that each time the same visitor is seen, |
| 246 | he's directed to the same server regardless of the number of available servers. |
| 247 | The information may be stored in the load balancer's memory, in which case it |
| 248 | may have to be replicated to other load balancers if it's not alone, or it may |
| 249 | be stored in the client's memory using various methods provided that the client |
| 250 | is able to present this information back with every request (cookie insertion, |
| 251 | redirection to a sub-domain, etc). This mechanism provides the extra benefit of |
| 252 | not having to rely on unstable or unevenly distributed information (such as the |
| 253 | source IP address). This is in fact the strongest reason to adopt a layer 7 |
| 254 | load balancer instead of a layer 4 one. |
| 255 | |
| 256 | In order to extract information such as a cookie, a host header field, a URL |
| 257 | or whatever, a load balancer may need to decrypt SSL/TLS traffic and even |
| 258 | possibly to reencrypt it when passing it to the server. This expensive task |
| 259 | explains why in some high-traffic infrastructures, sometimes there may be a |
| 260 | lot of load balancers. |
| 261 | |
| 262 | Since a layer 7 load balancer may perform a number of complex operations on the |
| 263 | traffic (decrypt, parse, modify, match cookies, decide what server to send to, |
| 264 | etc), it can definitely cause some trouble and will very commonly be accused of |
| 265 | being responsible for a lot of trouble that it only revealed. Often it will be |
| 266 | discovered that servers are unstable and periodically go up and down, or for |
| 267 | web servers, that they deliver pages with some hard-coded links forcing the |
| 268 | clients to connect directly to one specific server without passing via the load |
| 269 | balancer, or that they take ages to respond under high load causing timeouts. |
| 270 | That's why logging is an extremely important aspect of layer 7 load balancing. |
| 271 | Once a trouble is reported, it is important to figure if the load balancer took |
| 272 | a wrong decision and if so why so that it doesn't happen anymore. |
| 273 | |
| 274 | |
| 275 | 3. Introduction to HAProxy |
| 276 | -------------------------- |
| 277 | |
| 278 | HAProxy is written "HAProxy" to designate the product, "haproxy" to designate |
| 279 | the executable program, software package or a process, though both are commonly |
| 280 | used for both purposes, and is pronounced H-A-Proxy. Very early it used to stand |
| 281 | for "high availability proxy" and the name was written in two separate words, |
| 282 | though by now it means nothing else than "HAProxy". |
| 283 | |
| 284 | |
| 285 | 3.1. What HAProxy is and is not |
| 286 | ------------------------------- |
| 287 | |
| 288 | HAProxy is : |
| 289 | |
| 290 | - a TCP proxy : it can accept a TCP connection from a listening socket, |
| 291 | connect to a server and attach these sockets together allowing traffic to |
| 292 | flow in both directions; |
| 293 | |
| 294 | - an HTTP reverse-proxy (called a "gateway" in HTTP terminology) : it presents |
| 295 | itself as a server, receives HTTP requests over connections accepted on a |
| 296 | listening TCP socket, and passes the requests from these connections to |
| 297 | servers using different connections. |
| 298 | |
| 299 | - an SSL terminator / initiator / offloader : SSL/TLS may be used on the |
| 300 | connection coming from the client, on the connection going to the server, |
| 301 | or even on both connections. |
| 302 | |
| 303 | - a TCP normalizer : since connections are locally terminated by the operating |
| 304 | system, there is no relation between both sides, so abnormal traffic such as |
| 305 | invalid packets, flag combinations, window advertisements, sequence numbers, |
| 306 | incomplete connections (SYN floods), or so will not be passed to the other |
| 307 | side. This protects fragile TCP stacks from protocol attacks, and also |
| 308 | allows to optimize the connection parameters with the client without having |
| 309 | to modify the servers' TCP stack settings. |
| 310 | |
| 311 | - an HTTP normalizer : when configured to process HTTP traffic, only valid |
| 312 | complete requests are passed. This protects against a lot of protocol-based |
| 313 | attacks. Additionally, protocol deviations for which there is a tolerance |
| 314 | in the specification are fixed so that they don't cause problem on the |
| 315 | servers (eg: multiple-line headers). |
| 316 | |
| 317 | - an HTTP fixing tool : it can modify / fix / add / remove / rewrite the URL |
| 318 | or any request or response header. This helps fixing interoperability issues |
| 319 | in complex environments. |
| 320 | |
| 321 | - a content-based switch : it can consider any element from the request to |
| 322 | decide what server to pass the request or connection to. Thus it is possible |
| 323 | to handle multiple protocols over a same port (eg: http, https, ssh). |
| 324 | |
| 325 | - a server load balancer : it can load balance TCP connections and HTTP |
| 326 | requests. In TCP mode, load balancing decisions are taken for the whole |
| 327 | connection. In HTTP mode, decisions are taken per request. |
| 328 | |
| 329 | - a traffic regulator : it can apply some rate limiting at various points, |
| 330 | protect the servers against overloading, adjust traffic priorities based on |
| 331 | the contents, and even pass such information to lower layers and outer |
| 332 | network components by marking packets. |
| 333 | |
| 334 | - a protection against DDoS and service abuse : it can maintain a wide number |
| 335 | of statistics per IP address, URL, cookie, etc and detect when an abuse is |
| 336 | happening, then take action (slow down the offenders, block them, send them |
| 337 | to outdated contents, etc). |
| 338 | |
| 339 | - an observation point for network troubleshooting : due to the precision of |
| 340 | the information reported in logs, it is often used to narrow down some |
| 341 | network-related issues. |
| 342 | |
| 343 | - an HTTP compression offloader : it can compress responses which were not |
| 344 | compressed by the server, thus reducing the page load time for clients with |
| 345 | poor connectivity or using high-latency, mobile networks. |
| 346 | |
| 347 | HAProxy is not : |
| 348 | |
| 349 | - an explicit HTTP proxy, ie, the proxy that browsers use to reach the |
| 350 | internet. There are excellent open-source software dedicated for this task, |
| 351 | such as Squid. However HAProxy can be installed in front of such a proxy to |
| 352 | provide load balancing and high availability. |
| 353 | |
| 354 | - a caching proxy : it will return as-is the contents its received from the |
| 355 | server and will not interfere with any caching policy. There are excellent |
| 356 | open-source software for this task such as Varnish. HAProxy can be installed |
| 357 | in front of such a cache to provide SSL offloading, and scalability through |
| 358 | smart load balancing. |
| 359 | |
| 360 | - a data scrubber : it will not modify the body of requests nor responses. |
| 361 | |
| 362 | - a web server : during startup, it isolates itself inside a chroot jail and |
| 363 | drops its privileges, so that it will not perform any single file-system |
| 364 | access once started. As such it cannot be turned into a web server. There |
| 365 | are excellent open-source software for this such as Apache or Nginx, and |
| 366 | HAProxy can be installed in front of them to provide load balancing and |
| 367 | high availability. |
| 368 | |
| 369 | - a packet-based load balancer : it will not see IP packets nor UDP datagrams, |
| 370 | will not perform NAT or even less DSR. These are tasks for lower layers. |
| 371 | Some kernel-based components such as IPVS (Linux Virtual Server) already do |
| 372 | this pretty well and complement perfectly with HAProxy. |
| 373 | |
| 374 | |
| 375 | 3.2. How HAProxy works |
| 376 | ---------------------- |
| 377 | |
| 378 | HAProxy is a single-threaded, event-driven, non-blocking engine combining a very |
| 379 | fast I/O layer with a priority-based scheduler. As it is designed with a data |
| 380 | forwarding goal in mind, its architecture is optimized to move data as fast as |
| 381 | possible with the least possible operations. As such it implements a layered |
| 382 | model offering bypass mechanisms at each level ensuring data don't reach higher |
| 383 | levels when not needed. Most of the processing is performed in the kernel, and |
| 384 | HAProxy does its best to help the kernel do the work as fast as possible by |
| 385 | giving some hints or by avoiding certain operation when it guesses they could |
| 386 | be grouped later. As a result, typical figures show 15% of the processing time |
| 387 | spent in HAProxy versus 85% in the kernel in TCP or HTTP close mode, and about |
| 388 | 30% for HAProxy versus 70% for the kernel in HTTP keep-alive mode. |
| 389 | |
| 390 | A single process can run many proxy instances; configurations as large as |
| 391 | 300000 distinct proxies in a single process were reported to run fine. Thus |
| 392 | there is usually no need to start more than one process for all instances. |
| 393 | |
| 394 | It is possible to make HAProxy run over multiple processes, but it comes with |
| 395 | a few limitations. In general it doesn't make sense in HTTP close or TCP modes |
| 396 | because the kernel-side doesn't scale very well with some operations such as |
| 397 | connect(). It scales pretty well for HTTP keep-alive mode but the performance |
| 398 | that can be achieved out of a single process generaly outperforms common needs |
| 399 | by an order of magnitude. It does however make sense when used as an SSL |
| 400 | offloader, and this feature is well supported in multi-process mode. |
| 401 | |
| 402 | HAProxy only requires the haproxy executable and a configuration file to run. |
| 403 | For logging it is highly recommended to have a properly configured syslog daemon |
| 404 | and log rotations in place. The configuration files are parsed before starting, |
| 405 | then HAProxy tries to bind all listening sockets, and refuses to start if |
| 406 | anything fails. Past this point it cannot fail anymore. This means that there |
| 407 | are no runtime failures and that if it accepts to start, it will work until it |
| 408 | is stopped. |
| 409 | |
| 410 | Once HAProxy is started, it does exactly 3 things : |
| 411 | |
| 412 | - process incoming connections; |
| 413 | |
| 414 | - periodically check the servers' status (known as health checks); |
| 415 | |
| 416 | - exchange information with other haproxy nodes. |
| 417 | |
| 418 | Processing incoming connections is by far the most complex task as it depends |
| 419 | on a lot of configuration possibilities, but it can be summarized as the 9 steps |
| 420 | below : |
| 421 | |
| 422 | - accept incoming connections from listening sockets that belong to a |
| 423 | configuration entity known as a "frontend", which references one or multiple |
| 424 | listening addresses; |
| 425 | |
| 426 | - apply the frontend-specific processing rules to these connections that may |
| 427 | result in blocking them, modifying some headers, or intercepting them to |
| 428 | execute some internal applets such as the statistics page or the CLI; |
| 429 | |
| 430 | - pass these incoming connections to another configuration entity representing |
| 431 | a server farm known as a "backend", which contains the list of servers and |
| 432 | the load balancing strategy for this server farm; |
| 433 | |
| 434 | - apply the backend-specific processing rules to these connections; |
| 435 | |
| 436 | - decide which server to forward the connection to according to the load |
| 437 | balancing strategy; |
| 438 | |
| 439 | - apply the backend-specific processing rules to the response data; |
| 440 | |
| 441 | - apply the frontend-specific processing rules to the response data; |
| 442 | |
| 443 | - emit a log to report what happened in fine details; |
| 444 | |
| 445 | - in HTTP, loop back to the second step to wait for a new request, otherwise |
| 446 | close the connection. |
| 447 | |
| 448 | Frontends and backends are sometimes considered as half-proxies, since they only |
| 449 | look at one side of an end-to-end connection; the frontend only cares about the |
| 450 | clients while the backend only cares about the servers. HAProxy also supports |
| 451 | full proxies which are exactly the union of a frontend and a backend. When HTTP |
| 452 | processing is desired, the configuration will generally be split into frontends |
| 453 | and backends as they open a lot of possibilities since any frontend may pass a |
| 454 | connection to any backend. With TCP-only proxies, using frontends and backends |
| 455 | rarely provides a benefit and the configuration can be more readable with full |
| 456 | proxies. |
| 457 | |
| 458 | |
| 459 | 3.3. Basic features |
| 460 | ------------------- |
| 461 | |
| 462 | This section will enumerate a number of features that HAProxy implements, some |
| 463 | of which are generally expected from any modern load balancer, and some of |
| 464 | which are a direct benefit of HAProxy's architecture. More advanced features |
| 465 | will be detailed in the next section. |
| 466 | |
| 467 | |
| 468 | 3.3.1. Basic features : Proxying |
| 469 | -------------------------------- |
| 470 | |
| 471 | Proxying is the action of transferring data between a client and a server over |
| 472 | two independant connections. The following basic features are supported by |
| 473 | HAProxy regarding proxying and connection management : |
| 474 | |
| 475 | - Provide the server with a clean connection to protect them against any |
| 476 | client-side defect or attack; |
| 477 | |
| 478 | - Listen to multiple IP address and/or ports, even port ranges; |
| 479 | |
| 480 | - Transparent accept : intercept traffic targetting any arbitrary IP address |
| 481 | that doesn't even belong to the local system; |
| 482 | |
| 483 | - Server port doesn't need to be related to listening port, and may even be |
| 484 | translated by a fixed offset (useful with ranges); |
| 485 | |
| 486 | - Transparent connect : spoof the client's (or any) IP address if needed |
| 487 | when connecting to the server; |
| 488 | |
| 489 | - Provide a reliable return IP address to the servers in multi-site LBs; |
| 490 | |
| 491 | - Offload the server thanks to buffers and possibly short-lived connections |
| 492 | to reduce their concurrent connection count and their memory footprint; |
| 493 | |
| 494 | - Optimize TCP stacks (eg: SACK), congestion control, and reduce RTT impacts; |
| 495 | |
| 496 | - Support different protocol families on both sides (eg: IPv4/IPv6/Unix); |
| 497 | |
| 498 | - Timeout enforcement : HAProxy supports multiple levels of timeouts depending |
| 499 | on the stage the connection is, so that a dead client or server, or an |
| 500 | attacker cannot be granted resources for too long; |
| 501 | |
| 502 | - Protocol validation: HTTP, SSL, or payload are inspected and invalid |
| 503 | protocol elements are rejected, unless instructed to accept them anyway; |
| 504 | |
| 505 | - Policy enforcement : ensure that only what is allowed may be forwarded; |
| 506 | |
| 507 | - Both incoming and outgoing connections may be limited to certain network |
| 508 | namespaces (Linux only), making it easy to build a cross-container, |
| 509 | multi-tenant load balancer; |
| 510 | |
| 511 | - PROXY protocol presents the client's IP address to the server even for |
| 512 | non-HTTP traffic. This is an HAProxy extension that was adopted by a number |
| 513 | of third-party products by now, at least these ones at the time of writing : |
| 514 | - client : haproxy, stud, stunnel, exaproxy, ELB, squid |
| 515 | - server : haproxy, stud, postfix, exim, nginx, squid, node.js, varnish |
| 516 | |
| 517 | |
| 518 | 3.3.2. Basic features : SSL |
| 519 | --------------------------- |
| 520 | |
| 521 | HAProxy's SSL stack is recognized as one of the most featureful according to |
| 522 | Google's engineers (http://istlsfastyet.com/). The most commonly used features |
| 523 | making it quite complete are : |
| 524 | |
| 525 | - SNI-based multi-hosting with no limit on sites count and focus on |
| 526 | performance. At least one deployment is known for running 50000 domains |
| 527 | with their respective certificates; |
| 528 | |
| 529 | - support for wildcard certificates reduces the need for many certificates ; |
| 530 | |
| 531 | - certificate-based client authentication with configurable policies on |
| 532 | failure to present a valid certificate. This allows to present a different |
| 533 | server farm to regenerate the client certificate for example; |
| 534 | |
| 535 | - authentication of the backend server ensures the backend server is the real |
| 536 | one and not a man in the middle; |
| 537 | |
| 538 | - authentication with the backend server lets the backend server it's really |
| 539 | the expected haproxy node that is connecting to it; |
| 540 | |
| 541 | - TLS NPN and ALPN extensions make it possible to reliably offload SPDY/HTTP2 |
| 542 | connections and pass them in clear text to backend servers; |
| 543 | |
| 544 | - OCSP stapling further reduces first page load time by delivering inline an |
| 545 | OCSP response when the client requests a Certificate Status Request; |
| 546 | |
| 547 | - Dynamic record sizing provides both high performance and low latency, and |
| 548 | significantly reduces page load time by letting the browser start to fetch |
| 549 | new objects while packets are still in flight; |
| 550 | |
| 551 | - permanent access to all relevant SSL/TLS layer information for logging, |
| 552 | access control, reporting etc... These elements can be embedded into HTTP |
| 553 | header or even as a PROXY protocol extension so that the offloaded server |
| 554 | gets all the information it would have had if it performed the SSL |
| 555 | termination itself. |
| 556 | |
| 557 | - Detect, log and block certain known attacks even on vulnerable SSL libs, |
| 558 | such as the Heartbleed attack affecting certain versions of OpenSSL. |
| 559 | |
| 560 | |
| 561 | 3.3.3. Basic features : Monitoring |
| 562 | ---------------------------------- |
| 563 | |
| 564 | HAProxy focuses a lot on availability. As such it cares about servers state, |
| 565 | and about reporting its own state to other network components : |
| 566 | |
| 567 | - Servers state is continuously monitored using per-server parameters. This |
| 568 | ensures the path to the server is operational for regular traffic; |
| 569 | |
| 570 | - Health checks support two hysteresis for up and down transitions in order |
| 571 | to protect against state flapping; |
| 572 | |
| 573 | - Checks can be sent to a different address/port/protocol : this makes it |
| 574 | easy to check a single service that is considered representative of multiple |
| 575 | ones, for example the HTTPS port for an HTTP+HTTPS server. |
| 576 | |
| 577 | - Servers can track other servers and go down simultaneously : this ensures |
| 578 | that servers hosting multiple services can fail atomically and that noone |
| 579 | will be sent to a partially failed server; |
| 580 | |
| 581 | - Agents may be deployed on the server to monitor load and health : a server |
| 582 | may be interested in reporting its load, operational status, administrative |
| 583 | status independantly from what health checks can see. By running a simple |
| 584 | agent on the server, it's possible to consider the server's view of its own |
| 585 | health in addition to the health checks validating the whole path; |
| 586 | |
| 587 | - Various check methods are available : TCP connect, HTTP request, SMTP hello, |
| 588 | SSL hello, LDAP, SQL, Redis, send/expect scripts, all with/without SSL; |
| 589 | |
| 590 | - State change is notified in the logs and stats page with the failure reason |
| 591 | (eg: the HTTP response received at the moment the failure was detected); |
| 592 | |
| 593 | - Server state is also reported on the stats interface and can be used to take |
| 594 | routing decisions so that traffic may be sent to different farms depending |
| 595 | on their sizes and/or health (eg: loss of an inter-DC link); |
| 596 | |
| 597 | - HAProxy can use health check requests to pass information to the servers, |
| 598 | such as their names, weight, the number of other servers in the farm etc... |
| 599 | so that servers can adjust their response and decisions based on this |
| 600 | knowledge (eg: postpone backups to keep more CPU available); |
| 601 | |
| 602 | - Servers can use health checks to report more detailed state than just on/off |
| 603 | (eg: I would like to stop, please stop sending new visitors); |
| 604 | |
| 605 | - HAProxy itself can report its state to external components such as routers |
| 606 | or other load balancers, allowing to build very complete multi-path and |
| 607 | multi-layer infrastructures. |
| 608 | |
| 609 | |
| 610 | 3.3.4. Basic features : High availability |
| 611 | ----------------------------------------- |
| 612 | |
| 613 | Just like any serious load balancer, HAProxy cares a lot about availability to |
| 614 | ensure the best global service continuity : |
| 615 | |
| 616 | - Only valid servers are used ; the other ones are automatically evinced from |
| 617 | load balancing farms ; under certain conditions it is still possible to |
| 618 | force to use them though; |
| 619 | |
| 620 | - Support for a graceful shutdown so that it is possible to take servers out |
| 621 | of a farm without affecting any connection; |
| 622 | |
| 623 | - Backup servers are automatically used when active servers are down and |
| 624 | replace them so that sessions are not lost when possible. This also allows |
| 625 | to build multiple paths to reach the same server (eg: multiple interfaces); |
| 626 | |
| 627 | - Ability to return a global failed status for a farm when too many servers |
| 628 | are down. This, combined with the monitoring capabilities makes it possible |
| 629 | for an upstream component to choose a different LB node for a given service; |
| 630 | |
| 631 | - Stateless design makes it easy to build clusters : by design, HAProxy does |
| 632 | its best to ensure the highest service continuity without having to store |
| 633 | information that could be lost in the event of a failure. This ensures that |
| 634 | a takeover is the most seamless possible; |
| 635 | |
| 636 | - Integrates well with standard VRRP daemon keepalived : HAProxy easily tells |
| 637 | keepalived about its state and copes very will with floating virtual IP |
| 638 | addresses. Note: only use IP redundancy protocols (VRRP/CARP) over cluster- |
| 639 | based solutions (Heartbeat, ...) as they're the ones offering the fastest, |
| 640 | most seamless, and most reliable switchover. |
| 641 | |
| 642 | |
| 643 | 3.3.5. Basic features : Load balancing |
| 644 | -------------------------------------- |
| 645 | |
| 646 | HAProxy offers a fairly complete set of load balancing features, most of which |
| 647 | are unfortunately not available in a number of other load balancing products : |
| 648 | |
| 649 | - no less than 9 load balancing algorithms are supported, some of which apply |
| 650 | to input data to offer an infinite list of possibilities. The most common |
| 651 | ones are round-robin (for short connections, pick each server in turn), |
| 652 | leastconn (for long connections, pick the least recently used of the servers |
| 653 | with the lowest connection count), source (for SSL farms or terminal server |
| 654 | farms, the server directly depends on the client's source address), uri (for |
| 655 | HTTP caches, the server directly depends on the HTTP URI), hdr (the server |
| 656 | directly depends on the contents of a specific HTTP header field), first |
| 657 | (for short-lived virtual machines, all connections are packed on the |
| 658 | smallest possible subset of servers so that unused ones can be powered |
| 659 | down); |
| 660 | |
| 661 | - all algorithms above support per-server weights so that it is possible to |
| 662 | accommodate from different server generations in a farm, or direct a small |
| 663 | fraction of the traffic to specific servers (debug mode, running the next |
| 664 | version of the software, etc); |
| 665 | |
| 666 | - dynamic weights are supported for round-robin, leastconn and consistent |
| 667 | hashing ; this allows server weights to be modified on the fly from the CLI |
| 668 | or even by an agent running on the server; |
| 669 | |
| 670 | - slow-start is supported whenever a dynamic weight is supported; this allows |
| 671 | a server to progressively take the traffic. This is an important feature |
| 672 | for fragile application servers which require to compile classes at runtime |
| 673 | as well as cold caches which need to fill up before being run at full |
| 674 | throttle; |
| 675 | |
| 676 | - hashing can apply to various elements such as client's source address, URL |
| 677 | components, query string element, header field values, POST parameter, RDP |
| 678 | cookie; |
| 679 | |
| 680 | - consistent hashing protects server farms against massive redistribution when |
| 681 | adding or removing servers in a farm. That's very important in large cache |
| 682 | farms and it allows slow-start to be used to refill cold caches; |
| 683 | |
| 684 | - a number of internal metrics such as the number of connections per server, |
| 685 | per backend, the amount of available connection slots in a backend etc makes |
| 686 | it possible to build very advanced load balancing strategies. |
| 687 | |
| 688 | |
| 689 | 3.3.6. Basic features : Stickiness |
| 690 | ---------------------------------- |
| 691 | |
| 692 | Application load balancing would be useless without stickiness. HAProxy provides |
| 693 | a fairly comprehensive set of possibilities to maintain a visitor on the same |
| 694 | server even across various events such as server addition/removal, down/up |
| 695 | cycles, and some methods are designed to be resistant to the distance between |
| 696 | multiple load balancing nodes in that they don't require any replication : |
| 697 | |
| 698 | - stickiness information can be individually matched and learned from |
| 699 | different places if desired. For example a JSESSIONID cookie may be matched |
| 700 | both in a cookie and in the URL. Up to 8 parallel sources can be learned at |
| 701 | the same time and each of them may point to a different stick-table; |
| 702 | |
| 703 | - stickiness information can come from anything that can be seen within a |
| 704 | request or response, including source address, TCP payload offset and |
| 705 | length, HTTTP query string elements, header field values, cookies, and so |
| 706 | on... |
| 707 | |
| 708 | - stick-tables are replicated between all nodes in a multi-master fashion ; |
| 709 | |
| 710 | - commonly used elements such as SSL-ID or RDP cookies (for TSE farms) are |
| 711 | directly accessible to ease manipulation; |
| 712 | |
| 713 | - all sticking rules may be dynamically conditionned by ACLs; |
| 714 | |
| 715 | - it is possible to decide not to stick to certain servers, such as backup |
| 716 | servers, so that when the nominal server comes back, it automatically takes |
| 717 | the load back. This is often used in multi-path environments; |
| 718 | |
| 719 | - in HTTP it is often prefered not to learn anything and instead manipulate |
| 720 | a cookie dedicated to stickiness. For this, it's possible to detect, |
| 721 | rewrite, insert or prefix such a cookie to let the client remember what |
| 722 | server was assigned; |
| 723 | |
| 724 | - the server may decide to change or clean the stickiness cookie on logout, |
| 725 | so that leaving visitors are automatically unbound from the server; |
| 726 | |
| 727 | - using ACL-based rules it is also possible to selectively ignore or enforce |
| 728 | stickiness regardless of the server's state; combined with advanced health |
| 729 | checks, that helps admins verify that the server they're installing is up |
| 730 | and running before presenting it to the whole world; |
| 731 | |
| 732 | - an innovative mechanism to set a maximum idle time and duration on cookies |
| 733 | ensures that stickiness can be smoothly stopped on devices which are never |
| 734 | closed (smartphones, TVs, home appliances) without having to store them on |
| 735 | persistent storage; |
| 736 | |
| 737 | - multiple server entries may share the same stickiness keys so that |
| 738 | stickiness is not lost in multi-path environments when one path goes down; |
| 739 | |
| 740 | - soft-stop ensures that only users with stickiness information will continue |
| 741 | to reach the server they've been assigned to but no new users will go there. |
| 742 | |
| 743 | |
| 744 | 3.3.7. Basic features : Sampling and converting information |
| 745 | ----------------------------------------------------------- |
| 746 | |
| 747 | HAProxy supports information sampling using a wide set of "sample fetch |
| 748 | functions". The principle is to extract pieces of information known as samples, |
| 749 | for immediate use. This is used for stickiness, to build conditions, to produce |
| 750 | information in logs or to enrich HTTP headers. |
| 751 | |
| 752 | Samples can be fetched from various sources : |
| 753 | |
| 754 | - constants : integers, strings, IP addresses, binary blocks; |
| 755 | |
| 756 | - the process : date, environment variables, server/frontend/backend/process |
| 757 | state, byte/connection counts/rates, queue length, random generator, ... |
| 758 | |
| 759 | - variables : per-session, per-request, per-response variables; |
| 760 | |
| 761 | - the client connection : source and destination addresses and ports, and all |
| 762 | related statistics counters; |
| 763 | |
| 764 | - the SSL client session : protocol, version, algorithm, cipher, key size, |
| 765 | session ID, all client and server certificate fields, certificate serial, |
| 766 | SNI, ALPN, NPN, client support for certain extensions; |
| 767 | |
| 768 | - request and response buffers contents : arbitrary payload at offset/length, |
| 769 | data length, RDP cookie, decoding of SSL hello type, decoding of TLS SNI; |
| 770 | |
| 771 | - HTTP (request and response) : method, URI, path, query string arguments, |
| 772 | status code, headers values, positionnal header value, cookies, captures, |
| 773 | authentication, body elements; |
| 774 | |
| 775 | A sample may then pass through a number of operators known as "converters" to |
| 776 | experience some transformation. A converter consumes a sample and produces a |
| 777 | new one, possibly of a completely different type. For example, a converter may |
| 778 | be used to return only the integer length of the input string, or could turn a |
| 779 | string to upper case. Any arbitrary number of converters may be applied in |
| 780 | series to a sample before final use. Among all available sample converters, the |
| 781 | following ones are the most commonly used : |
| 782 | |
| 783 | - arithmetic and logic operators : they make it possible to perform advanced |
| 784 | computation on input data, such as computing ratios, percentages or simply |
| 785 | converting from one unit to another one; |
| 786 | |
| 787 | - IP address masks are useful when some addresses need to be grouped by larger |
| 788 | networks; |
| 789 | |
| 790 | - data representation : url-decode, base64, hex, JSON strings, hashing; |
| 791 | |
| 792 | - string conversion : extract substrings at fixed positions, fixed length, |
| 793 | extract specific fields around certain delimiters, extract certain words, |
| 794 | change case, apply regex-based substitution ; |
| 795 | |
| 796 | - date conversion : convert to http date format, convert local to UTC and |
| 797 | conversely, add or remove offset; |
| 798 | |
| 799 | - lookup an entry in a stick table to find statistics or assigned server; |
| 800 | |
| 801 | - map-based key-to-value conversion from a file (mostly used for geolocation). |
| 802 | |
| 803 | |
| 804 | 3.3.8. Basic features : Maps |
| 805 | ---------------------------- |
| 806 | |
| 807 | Maps are a powerful type of converter consisting in loading a two-columns file |
| 808 | into memory at boot time, then looking up each input sample from the first |
| 809 | column and either returning the corresponding pattern on the second column if |
| 810 | the entry was found, or returning a default value. The output information also |
| 811 | being a sample, it can in turn experience other transformations including other |
| 812 | map lookups. Maps are most commonly used to translate the client's IP address |
| 813 | to an AS number or country code since they support a longest match for network |
| 814 | addresses but they can be used for various other purposes. |
| 815 | |
| 816 | Part of their strength comes from being updatable on the fly either from the CLI |
| 817 | or from certain actions using other samples, making them capable of storing and |
| 818 | retrieving information between subsequent accesses. Another strength comes from |
| 819 | the binary tree based indexation which makes them extremely fast event when they |
| 820 | contain hundreds of thousands of entries, making geolocation very cheap and easy |
| 821 | to set up. |
| 822 | |
| 823 | |
| 824 | 3.3.9. Basic features : ACLs and conditions |
| 825 | ------------------------------------------- |
| 826 | |
| 827 | Most operations in HAProxy can be made conditional. Conditions are built by |
| 828 | combining multiple ACLs using logic operators (AND, OR, NOT). Each ACL is a |
| 829 | series of tests based on the following elements : |
| 830 | |
| 831 | - a sample fetch method to retrieve the element to test ; |
| 832 | |
| 833 | - an optional series of converters to transform the element ; |
| 834 | |
| 835 | - a list of patterns to match against ; |
| 836 | |
| 837 | - a matching method to indicate how to compare the patterns with the sample |
| 838 | |
| 839 | For example, the sample may be taken from the HTTP "Host" header, it could then |
| 840 | be converted to lower case, then matched against a number of regex patterns |
| 841 | using the regex matching method. |
| 842 | |
| 843 | Technically, ACLs are built on the same core as the maps, they share the exact |
| 844 | same internal structure, pattern matching methods and performance. The only real |
| 845 | difference is that instead of returning a sample, they only return "found" or |
| 846 | or "not found". In terms of usage, ACL patterns may be declared inline in the |
| 847 | configuration file and do not require their own file. ACLs may be named for ease |
| 848 | of use or to make configurations understandable. A named ACL may be declared |
| 849 | multiple times and it will evaluate all definitions in turn until one matches. |
| 850 | |
| 851 | About 13 different pattern matching methods are provided, among which IP address |
| 852 | mask, integer ranges, substrings, regex. They work like functions, and just like |
| 853 | with any programming language, only what is needed is evaluated, so when a |
| 854 | condition involving an OR is already true, next ones are not evaluated, and |
| 855 | similarly when a condition involving an AND is already false, the rest of the |
| 856 | condition is not evaluated. |
| 857 | |
| 858 | There is no practical limit to the number of declared ACLs, and a handful of |
| 859 | commonly used ones are provided. However experience has shown that setups using |
| 860 | a lot of named ACLs are quite hard to troubleshoot and that sometimes using |
| 861 | anynmous ACLs inline is easier as it requires less references out of the scope |
| 862 | being analysed. |
| 863 | |
| 864 | |
| 865 | 3.3.10. Basic features : Content switching |
| 866 | ------------------------------------------ |
| 867 | |
| 868 | HAProxy implements a mechanism known as content-based switching. The principle |
| 869 | is that a connection or request arrives on a frontend, then the information |
| 870 | carried with this request or connection are processed, and at this point it is |
| 871 | possible to write ACLs-based conditions making use of these information to |
| 872 | decide what backend will process the request. Thus the traffic is directed to |
| 873 | one backend or another based on the request's contents. The most common example |
| 874 | consists in using the Host header and/or elements from the path (sub-directories |
| 875 | or file-name extensions) to decide whether an HTTP request targets a static |
| 876 | object or the application, and to route static objects traffic to a backend made |
| 877 | of fast and light servers, and all the remaining traffic to a more complex |
| 878 | application server, thus constituting a fine-grained virtual hosting solution. |
| 879 | This is quite convenient to make multiple technologies coexist as a more global |
| 880 | solution. |
| 881 | |
| 882 | Another use case of content-switching consists in using different load balancing |
| 883 | algorithms depending on various criteria. A cache may use a URI hash while an |
| 884 | application would use round robin. |
| 885 | |
| 886 | Last but not least, it allows multiple customers to use a small share of a |
| 887 | common resource by enforcing per-backend (thus per-customer connection limits). |
| 888 | |
| 889 | Content switching rules scale very well, though their performance may depend on |
| 890 | the number and complexity of the ACLs in use. But it is also possible to write |
| 891 | dynamic content switching rules where a sample value directly turns into a |
| 892 | backend name and without making use of ACLs at all. Such configurations have |
| 893 | been reported to work fine at least with 300000 backends in production. |
| 894 | |
| 895 | |
| 896 | 3.3.11. Basic features : Stick-tables |
| 897 | ------------------------------------- |
| 898 | |
| 899 | Stick-tables are commonly used to store stickiness information, that is, to keep |
| 900 | a reference to the server a certain visitor was directed to. The key is then the |
| 901 | identifier associated with the visitor (its source address, the SSL ID of the |
| 902 | connection, an HTTP or RDP cookie, the customer number extracted from the URL or |
| 903 | from the payload, ...) and the stored value is then the server's identifier. |
| 904 | |
| 905 | Stick tables may use 3 different types of samples for their keys : integers, |
| 906 | strings and addresses. Only one stick-table may be referenced in a proxy, and it |
| 907 | is designated everywhere with the proxy name. Up to 8 key may be tracked in |
| 908 | parallel. The server identifier is committed during request or response |
| 909 | processing once both the key and the server are known. |
| 910 | |
| 911 | Stick-table contents may be replicated in active-active mode with other HAProxy |
| 912 | nodes known as "peers" as well as with the new process during a reload operation |
| 913 | so that all load balancing nodes share the same information and take the same |
| 914 | routing decision if a client's requests are spread over multiple nodes. |
| 915 | |
| 916 | Since stick-tables are indexed on what allows to recognize a client, they are |
| 917 | often also used to store extra information such as per-client statistics. The |
| 918 | extra statistics take some extra space and need to be explicitly declared. The |
| 919 | type of statistics that may be stored includes the input and output bandwidth, |
| 920 | the number of concurrent connections, the connection rate and count over a |
| 921 | period, the amount and frequency of errors, some specific tags and counters, |
| 922 | etc... In order to support keeping such information without being forced to |
| 923 | stick to a given server, a special "tracking" feature is implemented and allows |
| 924 | to track up to 3 simultaneous keys from different tables at the same time |
| 925 | regardless of stickiness rules. Each stored statistics may be searched, dumped |
| 926 | and cleared from the CLI and adds to the live troubleshooting capabilities. |
| 927 | |
| 928 | While this mechanism can be used to surclass a returning visitor or to adjust |
| 929 | the delivered quality of service depending on good or bad behaviour, it is |
| 930 | mostly used to fight against service abuse and more generally DDoS as it allows |
| 931 | to build complex models to detect certain bad behaviours at a high processing |
| 932 | speed. |
| 933 | |
| 934 | |
| 935 | 3.3.12. Basic features : Formated strings |
| 936 | ----------------------------------------- |
| 937 | |
| 938 | There are many places where HAProxy needs to manipulate character strings, such |
| 939 | as logs, redirects, header additions, and so on. In order to provide the |
| 940 | greatest flexibility, the notion of formated strings was introduced, initially |
| 941 | for logging purposes, which explains why it's still called "log-format". These |
| 942 | strings contain escape characters allowing to introduce various dynamic data |
| 943 | including variables and sample fetch expressions into strings, and even to |
| 944 | adjust the encoding while the result is being turned into a string (for example, |
| 945 | adding quotes). This provides a powerful way to build header contents or to |
| 946 | customize log lines. Additionally, in order to remain simple to build most |
| 947 | common strings, about 50 special tags are provided as shortcuts for information |
| 948 | commonly used in logs. |
| 949 | |
| 950 | |
| 951 | 3.3.13. Basic features : HTTP rewriting and redirection |
| 952 | ------------------------------------------------------- |
| 953 | |
| 954 | Installing a load balancer in front of an application that was never designed |
| 955 | for this can be a challenging task without the proper tools. One of the most |
| 956 | commonly requested operation in this case is to adjust requests and response |
| 957 | headers to make the load balancer appear as the origin server and to fix hard |
| 958 | coded information. This comes with changing the path in requests (which is |
| 959 | strongly advised against), modifying Host header field, modifying the Location |
| 960 | response header field for redirects, modifying the path and domain attribute |
| 961 | for cookies, and so on. It also happens that a number of servers are somewhat |
| 962 | verbose and tend to leak too much information in the response, making them more |
| 963 | vulnerable to targetted attacks. While it's theorically not the role of a load |
| 964 | balancer to clean this up, in practice it's located at the best place in the |
| 965 | infrastructure to guarantee that everything is cleaned up. |
| 966 | |
| 967 | Similarly, sometimes the load balancer will have to intercept some requests and |
| 968 | respond with a redirect to a new target URL. While some people tend to confuse |
| 969 | redirects and rewriting, these are two completely different concepts, since the |
| 970 | rewriting makes the client and the server see different things (and disagree on |
| 971 | the location of the page being visited) while redirects ask the client to visit |
| 972 | the new URL so that it sees the same location as the server. |
| 973 | |
| 974 | In order to do this, HAProxy supports various possibilities for rewriting and |
| 975 | redirect, among which : |
| 976 | |
| 977 | - regex-based URL and header rewriting in requests and responses. Regex are |
| 978 | the most commonly used tool to modify header values since they're easy to |
| 979 | manipulate and well understood; |
| 980 | |
| 981 | - headers may also be appended, deleted or replaced based on formated strings |
| 982 | so that it is possible to pass information there (eg: client side TLS |
| 983 | algorithm and cipher); |
| 984 | |
| 985 | - HTTP redirects can use any 3xx code to a relative, absolute, or completely |
| 986 | dynamic (formated string) URI; |
| 987 | |
| 988 | - HTTP redirects also support some extra options such as setting or clearing |
| 989 | a specific cookie, dropping the query string, appending a slash if missing, |
| 990 | and so on; |
| 991 | |
| 992 | - all operations support ACL-based conditions; |
| 993 | |
| 994 | |
| 995 | 3.3.14. Basic features : Server protection |
| 996 | ------------------------------------------ |
| 997 | |
| 998 | HAProxy does a lot to maximize service availability, and for this it deploys |
| 999 | large efforts to protect servers against overloading and attacks. The first |
| 1000 | and most important point is that only complete and valid requests are forwarded |
| 1001 | to the servers. The initial reason is that HAProxy needs to find the protocol |
| 1002 | elements it needs to stay synchronized with the byte stream, and the second |
| 1003 | reason is that until the request is complete, there is no way to know if some |
| 1004 | elements will change its semantics. The direct benefit from this is that servers |
| 1005 | are not exposed to invalid or incomplete requests. This is a very effective |
| 1006 | protection against slowloris attacks, which have almost no impact on HAProxy. |
| 1007 | |
| 1008 | Another important point is that HAProxy contains buffers to store requests and |
| 1009 | responses, and that by only sending a request to a server when it's complete and |
| 1010 | by reading the whole response very quickly from the local network, the server |
| 1011 | side connection is used for a very short time and this preserves server |
| 1012 | resources as much as possible. |
| 1013 | |
| 1014 | A direct extension to this is that HAProxy can artificially limit the number of |
| 1015 | concurrent connections or outstanding requests to a server, which guarantees |
| 1016 | that the server will never be overloaded even if it continuously runs at 100% of |
| 1017 | its capacity during traffic spikes. All excess requests will simply be queued to |
| 1018 | be processed when one slot is released. In the end, this huge resource savings |
| 1019 | most often ensures so much better server response times that it ends up actually |
| 1020 | being faster than by overloading the server. Queued requests may be redispatched |
| 1021 | to other servers, or even aborted in queue when the client aborts, which also |
| 1022 | protects the servers against the "reload effect", where each click on "reload" |
| 1023 | by a visitor on a slow-loading page usually induces a new request and maintains |
| 1024 | the server in an overloaded state. |
| 1025 | |
| 1026 | The slow-start mechanism also protects restarting servers against high traffic |
| 1027 | levels while they're still finalizing their startup or compiling some classes. |
| 1028 | |
| 1029 | Regarding the protocol-level protection, it is possible to relax the HTTP parser |
| 1030 | to accept non stardard-compliant but harmless requests or responses and even to |
| 1031 | fix them. This allows bogus applications to be accessible while a fix is being |
| 1032 | developped. In parallel, offending messages are completely captured with a |
| 1033 | detailed report that help developers spot the issue in the application. The most |
| 1034 | dangerous protocol violations are properly detected and dealt with and fixed. |
| 1035 | For example malformed requests or responses with two Content-length headers are |
| 1036 | either fixed if the values are exactly the same, or rejected if they differ, |
| 1037 | since it becomes a security problem. Protocol inspection is not limited to HTTP, |
| 1038 | it is also available for other protocols like TLS or RDP. |
| 1039 | |
| 1040 | When a protocol violation or attack is detected, there are various options to |
| 1041 | respond to the user, such as returning the common "HTTP 400 bad request", |
| 1042 | closing the connection with a TCP reset, faking an error after a long delay |
| 1043 | ("tarpit") to confuse the attacker. All of these contribute to protecting the |
| 1044 | servers by discouraging the offending client from pursuing an attack that |
| 1045 | becomes very expensive to maintain. |
| 1046 | |
| 1047 | HAProxy also proposes some more advanced options to protect against accidental |
| 1048 | data leaks and session crossing. Not only it can log suspicious server responses |
| 1049 | but it will also log and optionally block a response which might affect a given |
| 1050 | visitors' confidentiality. One such example is a cacheable cookie appearing in a |
| 1051 | cacheable response and which may result in an intermediary cache to deliver it |
| 1052 | to another visitor, causing an accidental session sharing. |
| 1053 | |
| 1054 | |
| 1055 | 3.3.15. Basic features : Logging |
| 1056 | -------------------------------- |
| 1057 | |
| 1058 | Logging is an extremely important feature for a load balancer, first because a |
| 1059 | load balancer is often accused of the trouble it reveals, and second because it |
| 1060 | is placed at a critical point in an infrastructure where all normal and abnormal |
| 1061 | activity needs to be analysed and correlated with other components. |
| 1062 | |
| 1063 | HAProxy provides very detailed logs, with millisecond accuracy and the exact |
| 1064 | connection accept time that can be searched in firewalls logs (eg: for NAT |
| 1065 | correlation). By default, TCP and HTTP logs are quite detailed an contain |
| 1066 | everything needed for troubleshooting, such as source IP address and port, |
| 1067 | frontend, backend, server, timers (request receipt duration, queue duration, |
| 1068 | connection setup time, response headers time, data transfer time), global |
| 1069 | process state, connection counts, queue status, retries count, detailed |
| 1070 | stickiness actions and disconnect reasons, header captures with a safe output |
| 1071 | encoding. It is then possible to extend or replace this format to include any |
| 1072 | sampled data, variables, captures, resulting in very detailed information. For |
| 1073 | example it is possible to log the number cumulated requests for this client or |
| 1074 | the number of different URLs for the client. |
| 1075 | |
| 1076 | The log level may be adjusted per request using standard ACLs, so it is possible |
| 1077 | to automatically silent some logs considered as pollution and instead raise |
| 1078 | warnings when some abnormal behaviour happen for a small part of the traffic |
| 1079 | (eg: too many URLs or HTTP errors for a source address). Administrative logs are |
| 1080 | also emitted with their own levels to inform about the loss or recovery of a |
| 1081 | server for example. |
| 1082 | |
| 1083 | Each frontend and backend may use multiple independant log outputs, which eases |
| 1084 | multi-tenancy. Logs are preferably sent over UDP, maybe JSON-encoded, and are |
| 1085 | truncated after a configurable line length in order to guarantee delivery. |
| 1086 | |
| 1087 | |
| 1088 | 3.3.16. Basic features : Statistics |
| 1089 | ----------------------------------- |
| 1090 | |
| 1091 | HAProxy provides a web-based statistics reporting interface with authentication, |
| 1092 | security levels and scopes. It is thus possible to provide each hosted customer |
| 1093 | with his own page showing only his own instances. This page can be located in a |
| 1094 | hidden URL part of the regular web site so that no new port needs to be opened. |
| 1095 | This page may also report the availability of other HAProxy nodes so that it is |
| 1096 | easy to spot if everything works as expected at a glance. The view is synthetic |
| 1097 | with a lot of details accessible (such as error causes, last access and last |
| 1098 | change duration, etc), which are also accessible as a CSV table that other tools |
| 1099 | may import to draw graphs. The page may self-refresh to be used as a monitoring |
| 1100 | page on a large display. In administration mode, the page also allows to change |
| 1101 | server state to ease maintenance operations. |
| 1102 | |
| 1103 | |
| 1104 | 3.4. Advanced features |
| 1105 | ---------------------- |
| 1106 | |
| 1107 | 3.4.1. Advanced features : Management |
| 1108 | ------------------------------------- |
| 1109 | |
| 1110 | HAProxy is designed to remain extremely stable and safe to manage in a regular |
| 1111 | production environment. It is provided as a single executable file which doesn't |
| 1112 | require any installation process. Multiple versions can easily coexist, meaning |
| 1113 | that it's possible (and recommended) to upgrade instances progressively by |
| 1114 | order of criticity instead of migrating all of them at once. Configuration files |
| 1115 | are easily versionned. Configuration checking is done off-line so it doesn't |
| 1116 | require to restart a service that will possibly fail. During configuration |
| 1117 | checks, a number of advanced mistakes may be detected (eg: for example, a rule |
| 1118 | hiding another one, or stickiness that will not work) and detailed warnings and |
| 1119 | configuration hints are proposed to fix them. Backwards configuration file |
| 1120 | compatibility goes very far away in time, with version 1.5 still fully |
| 1121 | supporting configurations for versions 1.1 written 13 years before, and 1.6 |
| 1122 | only dropping support for almost unused, obsolete keywords that can be done |
| 1123 | differently. The configuration and software upgrade mechanism is smooth and non |
| 1124 | disruptive in that it allows old and new processes to coexist on the system, |
| 1125 | each handling its own connections. System status, build options and library |
| 1126 | compatibility are reported on startup. |
| 1127 | |
| 1128 | Some advanced features allow an application administrator to smoothly stop a |
| 1129 | server, detect when there's no activity on it anymore, then take it off-line, |
| 1130 | stop it, upgrade it and ensure it doesn't take any traffic while being upgraded, |
| 1131 | then test it again through the normal path without opening it to the public, and |
| 1132 | all of this without touching HAProxy at all. This ensures that even complicated |
| 1133 | production operations may be done during opening hours with all technical |
| 1134 | resources available. |
| 1135 | |
| 1136 | The process tries to save resources as much as possible, uses memory pools to |
| 1137 | save on allocation time and limit memory fragmentation, releases payload buffers |
| 1138 | as soon as their contents are sent, and supports enforcing strong memory limits |
| 1139 | above which connections have to wait for a buffer to become available instead of |
| 1140 | allocating more memory. This system helps guarantee memory usage in certain |
| 1141 | strict environments. |
| 1142 | |
| 1143 | A command line interface (CLI) is available as a UNIX or TCP socket, to perform |
| 1144 | a number of operations and to retrieve troubleshooting information. Everything |
| 1145 | done on this socket doesn't require a configuration change, so it is mostly used |
| 1146 | for temporary changes. Using this interface it is possible to change a server's |
| 1147 | address, weight and status, to consult statistics and clear counters, dump and |
| 1148 | clear stickiness tables, possibly selectively by key criteria, dump and kill |
| 1149 | client-side and server-side connections, dump captured errors with a detailed |
| 1150 | analysis of the exact cause and location of the error, dump, add and remove |
| 1151 | entries from ACLs and maps, update TLS shared secrets, apply connection limits |
| 1152 | and rate limits on the fly to arbitrary frontends (useful in shared hosting |
| 1153 | environments), and disable a specific frontend to release a listening port |
| 1154 | (useful when daytime operations are forbidden and a fix is needed nonetheless). |
| 1155 | |
| 1156 | For environments where SNMP is mandatory, at least two agents exist, one is |
| 1157 | provided with the HAProxy sources and relies on the Net-SNMP perl module. |
| 1158 | Another one is provided with the commercial packages and doesn't require Perl. |
| 1159 | Both are roughly equivalent in terms of coverage. |
| 1160 | |
| 1161 | It is often recommended to install 4 utilities on the machine where HAProxy is |
| 1162 | deployed : |
| 1163 | |
| 1164 | - socat (in order to connect to the CLI, though certain forks of netcat can |
| 1165 | also do it to some extents); |
| 1166 | |
| 1167 | - halog from the latest HAProxy version : this is the log analysis tool, it |
| 1168 | parses native TCP and HTTP logs extremely fast (1 to 2 GB per second) and |
| 1169 | extracts useful information and statistics such as requests per URL, per |
| 1170 | source address, URLs sorted by response time or error rate, termination |
| 1171 | codes etc... It was designed to be deployed on the production servers to |
| 1172 | help troubleshoot live issues so it has to be there ready to be used; |
| 1173 | |
| 1174 | - tcpdump : this is highly recommended to take the network traces needed to |
| 1175 | troubleshoot an issue that was made visible in the logs. There is a moment |
| 1176 | where application and haproxy's analysis will diverge and the network traces |
| 1177 | are the only way to say who's right and who's wrong. It's also fairly common |
| 1178 | to detect bugs in network stacks and hypervisors thanks to tcpdump; |
| 1179 | |
| 1180 | - strace : it is tcpdump's companion. It will report what HAProxy really sees |
| 1181 | and will help sort out the issues the operating system is responsible for |
| 1182 | from the ones HAProxy is responsible for. Strace is often requested when a |
| 1183 | bug in HAProxy is suspected; |
| 1184 | |
| 1185 | |
| 1186 | 3.4.2. Advanced features : System-specific capabilities |
| 1187 | ------------------------------------------------------- |
| 1188 | |
| 1189 | Depending on the operating system HAProxy is deployed on, certain extra features |
| 1190 | may be available or needed. While it is supported on a number of platforms, |
| 1191 | HAProxy is primarily developped on Linux, which explains why some features are |
| 1192 | only available on this platform. |
| 1193 | |
| 1194 | The transparent bind and connect features, the support for binding connections |
| 1195 | to a specific network interface, as well as the ability to bind multiple |
| 1196 | processes to the same IP address and ports are only available on Linux and BSD |
| 1197 | systems, though only Linux performs a kernel-side load balancing of the incoming |
| 1198 | requests between the available processes. |
| 1199 | |
| 1200 | On Linux, there are also a number of extra features and optimizations including |
| 1201 | support for network namespaces (also known as "containers") allowing HAProxy to |
| 1202 | be a gateway between all containers, the ability to set the MSS, Netfilter marks |
| 1203 | and IP TOS field on the client side connection, support for TCP FastOpen on the |
| 1204 | listening side, TCP user timeouts to let the kernel quickly kill connections |
| 1205 | when it detects the client has disappeared before the configured timeouts, TCP |
| 1206 | splicing to let the kernel forward data between the two sides of a connections |
| 1207 | thus avoiding multiple memory copies, the ability to enable the "defer-accept" |
| 1208 | bind option to only get notified of an incoming connection once data become |
| 1209 | available in the kernel buffers, and the ability to send the request with the |
| 1210 | ACK confirming a connect (sometimes called "biggy-back") which is enabled with |
| 1211 | the "tcp-smart-connect" option. On Linux, HAProxy also takes great care of |
| 1212 | manipulating the TCP delayed ACKs to save as many packets as possible on the |
| 1213 | network. |
| 1214 | |
| 1215 | Some systems have an unreliable clock which jumps back and forth in the past |
| 1216 | and in the future. This used to happen with some NUMA systems where multiple |
| 1217 | processors didn't see the exact same time of day, and recently it became more |
| 1218 | common in virtualized environments where the virtual clock has no relation with |
| 1219 | the real clock, resulting in huge time jumps (sometimes up to 30 seconds have |
| 1220 | been observed). This causes a lot of trouble with respect to timeout enforcement |
| 1221 | in general. Due to this flaw of these systems, HAProxy maintains its own |
| 1222 | monotonic clock which is based on the system's clock but where drift is measured |
| 1223 | and compensated for. This ensures that even with a very bad system clock, timers |
| 1224 | remain reasonably accurate and timeouts continue to work. Note that this problem |
| 1225 | affects all the software running on such systems and is not specific to HAProxy. |
| 1226 | The common effects are spurious timeouts or application freezes. Thus if this |
| 1227 | behaviour is detected on a system, it must be fixed, regardless of the fact that |
| 1228 | HAProxy protects itself against it. |
| 1229 | |
| 1230 | |
| 1231 | 3.4.3. Advanced features : Scripting |
| 1232 | ------------------------------------ |
| 1233 | |
| 1234 | HAProxy can be built with support for the Lua embedded language, which opens a |
| 1235 | wide area of new possibilities related to complex manipulation of requests or |
| 1236 | responses, routing decisions, statistics processing and so on. Using Lua it is |
| 1237 | even possible to establish parallel connections to other servers to exchange |
| 1238 | information. This way it becomes possible (though complex) to develop an |
| 1239 | authentication system for example. Please refer to the documentation in the file |
| 1240 | "doc/lua-api/index.rst" for more information on how to use Lua. |
| 1241 | |
| 1242 | |
| 1243 | 3.5. Sizing |
| 1244 | ----------- |
| 1245 | |
| 1246 | Typical CPU usage figures show 15% of the processing time spent in HAProxy |
| 1247 | versus 85% in the kernel in TCP or HTTP close mode, and about 30% for HAProxy |
| 1248 | versus 70% for the kernel in HTTP keep-alive mode. This means that the operating |
| 1249 | system and its tuning have a strong impact on the global performance. |
| 1250 | |
| 1251 | Usages vary a lot between users, some focus on bandwidth, other ones on request |
| 1252 | rate, others on connection concurrency, others on SSL performance. this section |
| 1253 | aims at providing a few elements to help in this task. |
| 1254 | |
| 1255 | It is important to keep in mind that every operation comes with a cost, so each |
| 1256 | individual operation adds its overhead on top of the other ones, which may be |
| 1257 | negligible in certain circumstances, and which may dominate in other cases. |
| 1258 | |
| 1259 | When processing the requests from a connection, we can say that : |
| 1260 | |
| 1261 | - forwarding data costs less than parsing request or response headers; |
| 1262 | |
| 1263 | - parsing request or response headers cost less than establishing then closing |
| 1264 | a connection to a server; |
| 1265 | |
| 1266 | - establishing an closing a connection costs less than a TLS resume operation; |
| 1267 | |
| 1268 | - a TLS resume operation costs less than a full TLS handshake with a key |
| 1269 | computation; |
| 1270 | |
| 1271 | - an idle connection costs less CPU than a connection whose buffers hold data; |
| 1272 | |
| 1273 | - a TLS context costs even more memory than a connection with data; |
| 1274 | |
| 1275 | So in practice, it is cheaper to process payload bytes than header bytes, thus |
| 1276 | it is easier to achieve high network bandwidth with large objects (few requests |
| 1277 | per volume unit) than with small objects (many requests per volume unit). This |
| 1278 | explains why maximum bandwidth is always measured with large objects, while |
| 1279 | request rate or connection rates are measured with small objects. |
| 1280 | |
| 1281 | Some operations scale well on multiple process spread over multiple processors, |
| 1282 | and others don't scale as well. Network bandwidth doesn't scale very far because |
| 1283 | the CPU is rarely the bottleneck for large objects, it's mostly the network |
| 1284 | bandwidth and data busses to reach the network interfaces. The connection rate |
| 1285 | doesn't scale well over multiple processors due to a few locks in the system |
| 1286 | when dealing with the local ports table. The request rate over persistent |
| 1287 | connections scales very well as it doesn't involve much memory nor network |
| 1288 | bandwidth and doesn't require to access locked structures. TLS key computation |
| 1289 | scales very well as it's totally CPU-bound. TLS resume scales moderately well, |
| 1290 | but reaches its limits around 4 processes where the overhead of accessing the |
| 1291 | shared table offsets the small gains expected from more power. |
| 1292 | |
| 1293 | The performance numbers one can expect from a very well tuned system are in the |
| 1294 | following range. It is important to take them as orders of magnitude and to |
| 1295 | expect significant variations in any direction based on the processor, IRQ |
| 1296 | setting, memory type, network interface type, operating system tuning and so on. |
| 1297 | |
| 1298 | The following numbers were found on a Core i7 running at 3.7 GHz equiped with |
| 1299 | a dual-port 10 Gbps NICs running Linux kernel 3.10, HAProxy 1.6 and OpenSSL |
| 1300 | 1.0.2. HAProxy was running as a single process on a single dedicated CPU core, |
| 1301 | and two extra cores were dedicated to network interrupts : |
| 1302 | |
| 1303 | - 20 Gbps of maximum network bandwidth in clear text for objects 256 kB or |
| 1304 | higher, 10 Gbps for 41kB or higher; |
| 1305 | |
| 1306 | - 4.6 Gbps of TLS traffic using AES256-GCM cipher with large objects; |
| 1307 | |
| 1308 | - 83000 TCP connections per second from client to server; |
| 1309 | |
| 1310 | - 82000 HTTP connections per second from client to server; |
| 1311 | |
| 1312 | - 97000 HTTP requests per second in server-close mode (keep-alive with the |
| 1313 | client, close with the server); |
| 1314 | |
| 1315 | - 243000 HTTP requests per second in end-to-end keep-alive mode; |
| 1316 | |
| 1317 | - 300000 filtered TCP connections per second (anti-DDoS) |
| 1318 | |
| 1319 | - 160000 HTTPS requests per second in keep-alive mode over persistent TLS |
| 1320 | connections; |
| 1321 | |
| 1322 | - 13100 HTTPS requests per second using TLS resumed connections; |
| 1323 | |
| 1324 | - 1300 HTTPS connections per second using TLS connections renegociated with |
| 1325 | RSA2048; |
| 1326 | |
| 1327 | - 20000 concurrent saturated connections per GB of RAM, including the memory |
| 1328 | required for system buffers; it is possible to do better with careful tuning |
| 1329 | but this setting it easy to achieve. |
| 1330 | |
| 1331 | - about 8000 concurrent TLS connections (client-side only) per GB of RAM, |
| 1332 | including the memory required for system buffers; |
| 1333 | |
| 1334 | - about 5000 concurrent end-to-end TLS connections (both sides) per GB of |
| 1335 | RAM including the memory required for system buffers; |
| 1336 | |
| 1337 | Thus a good rule of thumb to keep in mind is that the request rate is divided |
| 1338 | by 10 between TLS keep-alive and TLS resume, and between TLS resume and TLS |
| 1339 | renegociation, while it's only divided by 3 between HTTP keep-alive and HTTP |
| 1340 | close. Another good rule of thumb is to remember that a high frequency core |
| 1341 | with AES instructions can do around 5 Gbps of AES-GCM per core. |
| 1342 | |
| 1343 | Having more core rarely helps (except for TLS) and is even counter-productive |
| 1344 | due to the lower frequency. In general a small number of high frequency cores |
| 1345 | is better. |
| 1346 | |
| 1347 | Another good rule of thumb is to consider that on the same server, HAProxy will |
| 1348 | be able to saturate : |
| 1349 | |
| 1350 | - about 5-10 static file servers or caching proxies; |
| 1351 | |
| 1352 | - about 100 anti-virus proxies; |
| 1353 | |
| 1354 | - and about 100 to 1000 application servers depending on the technology in use. |
| 1355 | |
| 1356 | |
| 1357 | 3.6. How to get HAProxy |
| 1358 | ----------------------- |
| 1359 | |
| 1360 | HAProxy is an opensource project covered by the GPLv2 license, meaning that |
| 1361 | everyone is allowed to redistribute it provided that access to the sources is |
| 1362 | also provided upon request, especially if any modifications were made. |
| 1363 | |
| 1364 | HAProxy evolves as a main development branch called "master" or "mainline", from |
| 1365 | which new branches are derived once the code is considered stable. A lot of web |
| 1366 | sites run some development branches in production on a voluntarily basis, either |
| 1367 | to participate to the project or because they need a bleeding edge feature, and |
| 1368 | their feedback is highly valuable to fix bugs and judge the overall quality and |
| 1369 | stability of the version being developped. |
| 1370 | |
| 1371 | The new branches that are created when the code is stable enough constitute a |
| 1372 | stable version and are generally maintained for several years, so that there is |
| 1373 | no emergency to migrate to a newer branch even when you're not on the latest. |
| 1374 | Once a stable branch is issued, it may only receive bug fixes, and very rarely |
| 1375 | minor feature updates when that makes users' life easier. All fixes that go into |
| 1376 | a stable branch necessarily come from the master branch. This guarantees that no |
| 1377 | fix will be lost after an upgrade. For this reason, if you fix a bug, please |
| 1378 | make the patch against the master branch, not the stable branch. You may even |
| 1379 | discover it was already fixed. This process also ensures that regressions in a |
| 1380 | stable branch are extremely rare, so there is never any excuse for not upgrading |
| 1381 | to the latest version in your current branch. |
| 1382 | |
| 1383 | Branches are numberred with two digits delimited with a dot, such as "1.6". A |
| 1384 | complete version includes one or two sub-version numbers indicating the level of |
| 1385 | fix. For example, version 1.5.14 is the 14th fix release in branch 1.5 after |
| 1386 | version 1.5.0 was issued. It contains 126 fixes for individual bugs, 24 updates |
| 1387 | on the documentation, and 75 other backported patches, most of which were needed |
| 1388 | to fix the aforementionned 126 bugs. An existing feature may never be modified |
| 1389 | nor removed in a stable branch, in order to guarantee that upgrades within the |
| 1390 | same branch will always be harmless. |
| 1391 | |
| 1392 | HAProxy is available from multiple sources, at different release rhythms : |
| 1393 | |
| 1394 | - The official community web site : http://www.haproxy.org/ : this site |
| 1395 | provides the sources of the latest development release, all stable releases, |
| 1396 | as well as nightly snapshots for each branch. The release cycle is not fast, |
| 1397 | several months between stable releases, or between development snapshots. |
| 1398 | Very old versions are still supported there. Everything is provided as |
| 1399 | sources only, so whatever comes from there needs to be rebuilt and/or |
| 1400 | repackaged; |
| 1401 | |
| 1402 | - A number of operating systems such as Linux distributions and BSD ports. |
| 1403 | These systems generally provide long-term maintained versions which do not |
| 1404 | always contain all the fixes from the official ones, but which at least |
| 1405 | contain the critical fixes. It often is a good option for most users who do |
| 1406 | not seek advanced configurations and just want to keep updates easy; |
| 1407 | |
| 1408 | - Commercial versions from http://www.haproxy.com/ : these are supported |
| 1409 | professional packages built for various operating systems or provided as |
| 1410 | appliances, based on the latest stable versions and including a number of |
| 1411 | features backported from the next release for which there is a strong |
| 1412 | demand. It is the best option for users seeking the latest features with |
| 1413 | the reliability of a stable branch, the fastest response time to fix bugs, |
| 1414 | or simply support contracts on top of an opensource product; |
| 1415 | |
| 1416 | |
| 1417 | In order to ensure that the version you're using is the latest one in your |
| 1418 | branch, you need to proceed this way : |
| 1419 | |
| 1420 | - verify which HAProxy executable you're running : some systems ship it by |
| 1421 | default and administrators install their versions somewhere else on the |
| 1422 | system, so it is important to verify in the startup scripts which one is |
| 1423 | used; |
| 1424 | |
| 1425 | - determine which source your HAProxy version comes from. For this, it's |
| 1426 | generally sufficient to type "haproxy -v". A development version will |
| 1427 | appear like this, with the "dev" word after the branch number : |
| 1428 | |
| 1429 | HA-Proxy version 1.6-dev3-385ecc-68 2015/08/18 |
| 1430 | |
| 1431 | A stable version will appear like this, as well as unmodified stable |
| 1432 | versions provided by operating system vendors : |
| 1433 | |
| 1434 | HA-Proxy version 1.5.14 2015/07/02 |
| 1435 | |
| 1436 | And a nightly snapshot of a stable version will appear like this with an |
| 1437 | hexadecimal sequence after the version, and with the date of the snapshot |
| 1438 | instead of the date of the release : |
| 1439 | |
| 1440 | HA-Proxy version 1.5.14-e4766ba 2015/07/29 |
| 1441 | |
| 1442 | Any other format may indicate a system-specific package with its own |
| 1443 | patch set. For example HAProxy Enterprise versions will appear with the |
| 1444 | following format (<branch>-<latest commit>-<revision>) : |
| 1445 | |
| 1446 | HA-Proxy version 1.5.0-994126-357 2015/07/02 |
| 1447 | |
| 1448 | - for system-specific packages, you have to check with your vendor's package |
| 1449 | repository or update system to ensure that your system is still supported, |
| 1450 | and that fixes are still provided for your branch. For community versions |
| 1451 | coming from haproxy.org, just visit the site, verify the status of your |
| 1452 | branch and compare the latest version with yours to see if you're on the |
| 1453 | latest one. If not you can upgrade. If your branch is not maintained |
| 1454 | anymore, you're definitely very late and will have to consider an upgrade |
| 1455 | to a more recent branch (carefully read the README when doing so). |
| 1456 | |
| 1457 | HAProxy will have to be updated according to the source it came from. Usually it |
| 1458 | follows the system vendor's way of upgrading a package. If it was taken from |
| 1459 | sources, please read the README file in the sources directory after extracting |
| 1460 | the sources and follow the instructions for your operating system. |
| 1461 | |
| 1462 | |
| 1463 | 4. Companion products and alternatives |
| 1464 | -------------------------------------- |
| 1465 | |
| 1466 | HAProxy integrates fairly well with certain products listed below, which is why |
| 1467 | they are mentionned here even if not directly related to HAProxy. |
| 1468 | |
| 1469 | |
| 1470 | 4.1. Apache HTTP server |
| 1471 | ----------------------- |
| 1472 | |
| 1473 | Apache is the de-facto standard HTTP server. It's a very complete and modular |
| 1474 | project supporting both file serving and dynamic contents. It can serve as a |
| 1475 | frontend for some application servers. In can even proxy requests and cache |
| 1476 | responses. In all of these use cases, a front load balancer is commonly needed. |
| 1477 | Apache can work in various modes, certain being heavier than other ones. Certain |
| 1478 | modules still require the heavier pre-forked model and will prevent Apache from |
| 1479 | scaling well with a high number of connections. In this case HAProxy can provide |
| 1480 | a tremendous help by enforcing the per-server connection limits to a safe value |
| 1481 | and will significantly speed up the server and preserve its resources that will |
| 1482 | be better used by the application. |
| 1483 | |
| 1484 | Apache can extract the client's address from the X-Forwarded-For header by using |
| 1485 | the "mod_rpaf" extension. HAProxy will automatically feed this header when |
| 1486 | "option forwardfor" is specified in its configuration. HAProxy may also offer a |
| 1487 | nice protection to Apache when exposed to the internet, where it will better |
| 1488 | resist to a wide number of types of DoS. |
| 1489 | |
| 1490 | |
| 1491 | 4.2. NGINX |
| 1492 | ---------- |
| 1493 | |
| 1494 | NGINX is the second de-facto standard HTTP server. Just like Apache, it covers a |
| 1495 | wide range of features. NGINX is built on a similar model as HAProxy so it has |
| 1496 | no problem dealing with tens of thousands of concurrent connections. When used |
| 1497 | as a gateway to some applications (eg: using the included PHP FPM), it can often |
| 1498 | be beneficial to set up some frontend connection limiting to reduce the load |
| 1499 | on the PHP application. HAProxy will clearly be useful there both as a regular |
| 1500 | load balancer and as the traffic regulator to speed up PHP by decongestionning |
| 1501 | it. Also since both products use very little CPU thanks to their event-driven |
| 1502 | architecture, it's often easy to install both of them on the same system. NGINX |
| 1503 | implements HAProxy's PROXY protocol, thus it is easy for HAProxy to pass the |
| 1504 | client's connection information to NGINX so that the application gets all the |
| 1505 | relevant information. Some benchmarks have also shown that for large static |
| 1506 | file serving, implementing consistent hash on HAProxy in front of NGINX can be |
| 1507 | beneficial by optimizing the OS' cache hit ratio, which is basically multiplied |
| 1508 | by the number of server nodes. |
| 1509 | |
| 1510 | |
| 1511 | 4.3. Varnish |
| 1512 | ------------ |
| 1513 | |
| 1514 | Varnish is a smart caching reverse-proxy, probably best described as a web |
| 1515 | application accelerator. Varnish doesn't implement SSL/TLS and wants to dedicate |
| 1516 | all of its CPU cycles to what it does best. Varnish also implements HAProxy's |
| 1517 | PROXY protocol so that HAProxy can very easily be deployed in front of Varnish |
| 1518 | as an SSL offloader as well as a load balancer and pass it all relevant client |
| 1519 | information. Also, Varnish naturally supports decompression from the cache when |
| 1520 | a server has provided a compressed object, but doesn't compress however. HAProxy |
| 1521 | can then be used to compress outgoing data when backend servers do not implement |
| 1522 | compression, though it's rarely a good idea to compress on the load balancer |
| 1523 | unless the traffic is low. |
| 1524 | |
| 1525 | When building large caching farms across multiple nodes, HAProxy can make use of |
| 1526 | consistent URL hashing to intelligently distribute the load to the caching nodes |
| 1527 | and avoid cache duplication, resulting in a total cache size which is the sum of |
| 1528 | all caching nodes. |
| 1529 | |
| 1530 | |
| 1531 | 4.4. Alternatives |
| 1532 | ----------------- |
| 1533 | |
| 1534 | Linux Virtual Server (LVS or IPVS) is the layer 4 load balancer included within |
| 1535 | the Linux kernel. It works at the packet level and handles TCP and UDP. In most |
| 1536 | cases it's more a complement than an alternative since it doesn't have layer 7 |
| 1537 | knowledge at all. |
| 1538 | |
| 1539 | Pound is another well-known load balancer. It's much simpler and has much less |
| 1540 | features than HAProxy but for many very basic setups both can be used. Its |
| 1541 | author has always focused on code auditability first and wants to maintain the |
| 1542 | set of features low. Its thread-based architecture scales less well with high |
| 1543 | connection counts, but it's a good product. |
| 1544 | |
| 1545 | Pen is a quite light load balancer. It supports SSL, maintains persistence using |
| 1546 | a fixed-size table of its clients' IP addresses. It supports a packet-oriented |
| 1547 | mode allowing it to support direct server return and UDP to some extents. It is |
| 1548 | meant for small loads (the persistence table only has 2048 entries). |
| 1549 | |
| 1550 | NGINX can do some load balancing to some extents, though it's clearly not its |
| 1551 | primary function. Production traffic is used to detect server failures, the |
| 1552 | load balancing algorithms are more limited, and the stickiness is very limited. |
| 1553 | But it can make sense in some simple deployment scenarios where it is already |
| 1554 | present. The good thing is that since it integrates very well with HAProxy, |
| 1555 | there's nothing wrong with adding HAProxy later when its limits have been faced. |
| 1556 | |
| 1557 | Varnish also does some load balancing of its backend servers and does support |
| 1558 | real health checks. It doesn't implement stickiness however, so just like with |
| 1559 | NGINX, as long as stickiness is not needed that can be enough to start with. |
| 1560 | And similarly, since HAProxy and Varnish integrate so well together, it's easy |
| 1561 | to add it later into the mix to complement the feature set. |
| 1562 | |