NGINX

How To Scale SSL with HAProxy and Nginx

SSL is CPU Intensive

If you haven’t already enabled SSL session caching, do that NOW. But what if you have many unique requests and your load balancer is maxing out it’s CPU? That was the case with WakaTime’s load balancer, because as you use the WakaTime plugins you are constantly making requests to our api saying you’re still working on a project. We had one load balancer terminating SSL in front of multiple app servers running ourFlask app. The Flask app servers handled the requests just fine, but the load balancer was maxing out all 16 cores negotiating SSL handshakes.

Proxying TCP insead of HTTP

The solution is to proxy TCP instead of HTTP. This way the load balancer no longer terminates SSL, but passes the TCP connection on to your app servers unmodified.

Let’s say you have two nginx app servers (10.0.0.11 and 10.0.0.12) and one haproxy load balancer (10.0.0.10). First, install haproxy. Then, edit /etc/haproxy/haproxy.cfg adding these lines:

frontend https-in
    bind *:443
    default_backend https-servers

backend https-servers
        mode tcp
        balance roundrobin
        server srv1 10.0.0.11:443
        server srv2 10.0.0.12:443

With an nginx config like:

server {
    server_name  example.com;
    root /opt/example/current/app;

    listen 443 ssl http2;

    ssl_certificate /etc/ssl/example/ssl.crt;
    ssl_certificate_key /etc/ssl/example/ssl.key;

    location / {
        include uwsgi_params;
        uwsgi_pass unix:/tmp/app.sock;
    }
}

This tells haproxy to setup a Layer 4 proxy to forward all TCP connections unmodified to the two nginx servers using roundrobin to balance the connections. The nginx app servers will share the load of negotiating SSL and parsing the HTTP requests.

load-balancing-haproxy-nginx-small

One catch though, your nginx app servers will see the requests coming from the IP address of your haproxy load balancer instead of the originating client. To fix this, enable Proxy Protocol to send the originating client’s IP address to your nginx app servers.

Forwarding the User’s Real IP using Proxy Protocol

Proxy Protocol forwards the originating client’s IP address from haproxy to nginx without having to modify the HTTP request headers. To enable Proxy Protocol in haproxy, add the send-proxy keyword to your /etc/haproxy/haproxy.cfg file:

frontend https-in
    bind *:443
    default_backend https-servers

backend https-servers
        mode tcp
        balance roundrobin
        server srv1 10.0.0.11:443 send-proxy
        server srv2 10.0.0.12:443 send-proxy

And tell nginx to receive the client’s real IP using proxy protocol and add it to the HTTP request:

server {
    server_name  example.com;
    root /opt/example/current/app;

    listen 443 ssl http2 proxy_protocol;

    set_real_ip_from 10.0.0.10/32;
    real_ip_header proxy_protocol;

    ssl_certificate /etc/ssl/example/ssl.crt;
    ssl_certificate_key /etc/ssl/example/ssl.key;

    location / {
        include uwsgi_params;
        uwsgi_pass unix:/tmp/app.sock;
    }
}

Notice how we told nginx to trust the IP address of your haproxy load balancer 10.0.0.10 to give us the client’s real IP. SSL is distributed among your two nginx app servers, and your nginx log files show the correct client IP address for each request.

Now you can scale to infinity!*

* Your haproxy process is still limited to a certain number of connections (ulimit).

Reference: How To Scale SSL with HAProxy and Nginx from our SCG partner Alan Hamlett at the Wakatime blog.
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

0 Comments
Inline Feedbacks
View all comments
Back to top button