Mercurial > hg > ngx_http_upstream_keepalive
view t/memcached-keepalive.t @ 11:15530a464dba
Keepalive: don't cache invalid connections.
1. Remember failed status, since peer.free() may be called more than once.
It's called twice if peer fails - once from ngx_http_upstream_next() with
NGX_PEER_FAILED set, and then from ngx_http_upstream_finalize_request()
without.
2. We shouldn't cache connection unless we aren't expecting anything from
upstream. For memcached this means either 404 response or 200 response with
all body read (body may not be read e.g. when serving HEAD request).
author | Maxim Dounin <mdounin@mdounin.ru> |
---|---|
date | Thu, 13 Nov 2008 19:36:23 +0300 |
parents | bef88ba0b378 |
children | 067ddc059ee0 |
line wrap: on
line source
#!/usr/bin/perl # (C) Maxim Dounin # Test for memcached with keepalive. ############################################################################### use warnings; use strict; use Test::More; use Test::Nginx; ############################################################################### select STDERR; $| = 1; select STDOUT; $| = 1; eval { require Cache::Memcached; }; plain(skip_all => 'Cache::Memcached not installed') if $@; my $t = Test::Nginx->new()->has('rewrite')->has_daemon('memcached')->plan(16) ->write_file_expand('nginx.conf', <<'EOF'); master_process off; daemon off; events { worker_connections 1024; } http { access_log off; client_body_temp_path %%TESTDIR%%/client_body_temp; fastcgi_temp_path %%TESTDIR%%/fastcgi_temp; proxy_temp_path %%TESTDIR%%/proxy_temp; upstream memd { server 127.0.0.1:8081; keepalive 1; } upstream memd2 { server 127.0.0.1:8081; server 127.0.0.1:8082; keepalive 1 single; } upstream memd3 { server 127.0.0.1:8081; server 127.0.0.1:8082; keepalive 1; } upstream memd4 { server 127.0.0.1:8081; server 127.0.0.1:8082; keepalive 10; } server { listen localhost:8080; server_name localhost; location / { set $memcached_key $uri; memcached_pass memd; } location /next { set $memcached_key $uri; memcached_next_upstream not_found; memcached_pass memd; } location /memd2 { set $memcached_key "/"; memcached_pass memd2; } location /memd3 { set $memcached_key "/"; memcached_pass memd3; } location /memd4 { set $memcached_key "/"; memcached_pass memd4; } } } EOF $t->run_daemon('memcached', '-l', '127.0.0.1', '-p', '8081'); $t->run_daemon('memcached', '-l', '127.0.0.1', '-p', '8082'); $t->run(); ############################################################################### my $memd1 = Cache::Memcached->new(servers => [ '127.0.0.1:8081' ]); my $memd2 = Cache::Memcached->new(servers => [ '127.0.0.1:8082' ]); $memd1->set('/', 'SEE-THIS'); $memd2->set('/', 'SEE-THIS'); $memd1->set('/big', 'X' x 1000000); my $total = $memd1->stats()->{total}->{total_connections}; like(http_get('/'), qr/SEE-THIS/, 'keepalive memcached request'); like(http_get('/notfound'), qr/404/, 'keepalive memcached not found'); like(http_get('/next'), qr/404/, 'keepalive not found with memcached_next_upstream'); like(http_get('/'), qr/SEE-THIS/, 'keepalive memcached request again'); like(http_get('/'), qr/SEE-THIS/, 'keepalive memcached request again'); like(http_get('/'), qr/SEE-THIS/, 'keepalive memcached request again'); is($memd1->stats()->{total}->{total_connections}, $total + 1, 'only one connection used'); # Since nginx doesn't read all data from connection in some situations (head # requests, post_action, errors writing to client) we have to close such # connections. Check if we really do close them. $total = $memd1->stats()->{total}->{total_connections}; unlike(http_head('/'), qr/SEE-THIS/, 'head request'); like(http_get('/'), qr/SEE-THIS/, 'get after head'); is($memd1->stats()->{total}->{total_connections}, $total + 1, 'head request closes connection'); $total = $memd1->stats()->{total}->{total_connections}; unlike(http_head('/big'), qr/XXX/, 'big head'); like(http_get('/'), qr/SEE-THIS/, 'get after big head'); is($memd1->stats()->{total}->{total_connections}, $total + 1, 'big head request closes connection'); # two backends with 'single' option - should establish only one connection $total = $memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}; http_get('/memd2'); http_get('/memd2'); http_get('/memd2'); is($memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}, $total + 1, 'only one connection with two backends and single'); $total = $memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}; # two backends without 'single' option and maximum number of cached # connections set to 1 - should establish new connection on each request http_get('/memd3'); http_get('/memd3'); http_get('/memd3'); is($memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}, $total + 3, '3 connections should be established'); # two backends without 'single' option and maximum number of cached # connections set to 10 - should establish only two connections (1 per backend) $total = $memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}; http_get('/memd4'); http_get('/memd4'); http_get('/memd4'); is($memd1->stats()->{total}->{total_connections} + $memd2->stats()->{total}->{total_connections}, $total + 2, 'connection per backend'); ###############################################################################