Merge pull request #4514 from alexmaret/4475-stickyness-mode
Added new affinity mode for maximum session stickyness.
This commit is contained in:
commit
f6c2f5fb97
18 changed files with 572 additions and 138 deletions
|
|
@ -6,7 +6,8 @@ local configuration = require("configuration")
|
|||
local round_robin = require("balancer.round_robin")
|
||||
local chash = require("balancer.chash")
|
||||
local chashsubset = require("balancer.chashsubset")
|
||||
local sticky = require("balancer.sticky")
|
||||
local sticky_balanced = require("balancer.sticky_balanced")
|
||||
local sticky_persistent = require("balancer.sticky_persistent")
|
||||
local ewma = require("balancer.ewma")
|
||||
|
||||
-- measured in seconds
|
||||
|
|
@ -19,7 +20,8 @@ local IMPLEMENTATIONS = {
|
|||
round_robin = round_robin,
|
||||
chash = chash,
|
||||
chashsubset = chashsubset,
|
||||
sticky = sticky,
|
||||
sticky_balanced = sticky_balanced,
|
||||
sticky_persistent = sticky_persistent,
|
||||
ewma = ewma,
|
||||
}
|
||||
|
||||
|
|
@ -30,7 +32,11 @@ local function get_implementation(backend)
|
|||
local name = backend["load-balance"] or DEFAULT_LB_ALG
|
||||
|
||||
if backend["sessionAffinityConfig"] and backend["sessionAffinityConfig"]["name"] == "cookie" then
|
||||
name = "sticky"
|
||||
if backend["sessionAffinityConfig"]["mode"] == 'persistent' then
|
||||
name = "sticky_persistent"
|
||||
else
|
||||
name = "sticky_balanced"
|
||||
end
|
||||
elseif backend["upstreamHashByConfig"] and backend["upstreamHashByConfig"]["upstream-hash-by"] then
|
||||
if backend["upstreamHashByConfig"]["upstream-hash-by-subset"] then
|
||||
name = "chashsubset"
|
||||
|
|
|
|||
|
|
@ -1,43 +1,38 @@
|
|||
local balancer_resty = require("balancer.resty")
|
||||
local resty_chash = require("resty.chash")
|
||||
local util = require("util")
|
||||
local ck = require("resty.cookie")
|
||||
local math = require("math")
|
||||
local ngx_balancer = require("ngx.balancer")
|
||||
local split = require("util.split")
|
||||
|
||||
local string_format = string.format
|
||||
local ngx_log = ngx.log
|
||||
local INFO = ngx.INFO
|
||||
|
||||
local _M = balancer_resty:new({ factory = resty_chash, name = "sticky" })
|
||||
local _M = balancer_resty:new()
|
||||
local DEFAULT_COOKIE_NAME = "route"
|
||||
|
||||
-- Consider the situation of N upstreams one of which is failing.
|
||||
-- Then the probability to obtain failing upstream after M iterations would be close to (1/N)**M.
|
||||
-- For the worst case (2 upstreams; 20 iterations) it would be ~10**(-6)
|
||||
-- which is much better then ~10**(-3) for 10 iterations.
|
||||
local MAX_UPSTREAM_CHECKS_COUNT = 20
|
||||
|
||||
function _M.cookie_name(self)
|
||||
return self.cookie_session_affinity.name or DEFAULT_COOKIE_NAME
|
||||
end
|
||||
|
||||
function _M.new(self, backend)
|
||||
local nodes = util.get_nodes(backend.endpoints)
|
||||
|
||||
function _M.new(self)
|
||||
local o = {
|
||||
instance = self.factory:new(nodes),
|
||||
traffic_shaping_policy = backend.trafficShapingPolicy,
|
||||
alternative_backends = backend.alternativeBackends,
|
||||
cookie_session_affinity = backend["sessionAffinityConfig"]["cookieSessionAffinity"]
|
||||
alternative_backends = nil,
|
||||
cookie_session_affinity = nil,
|
||||
traffic_shaping_policy = nil
|
||||
}
|
||||
|
||||
setmetatable(o, self)
|
||||
self.__index = self
|
||||
|
||||
return o
|
||||
end
|
||||
|
||||
local function set_cookie(self, value)
|
||||
function _M.get_cookie(self)
|
||||
local cookie, err = ck:new()
|
||||
if not cookie then
|
||||
ngx.log(ngx.ERR, err)
|
||||
end
|
||||
|
||||
return cookie:get(self:cookie_name())
|
||||
end
|
||||
|
||||
function _M.set_cookie(self, value)
|
||||
local cookie, err = ck:new()
|
||||
if not cookie then
|
||||
ngx.log(ngx.ERR, err)
|
||||
|
|
@ -86,22 +81,6 @@ local function get_failed_upstreams()
|
|||
return indexed_upstream_addrs
|
||||
end
|
||||
|
||||
local function pick_new_upstream(self)
|
||||
local failed_upstreams = get_failed_upstreams()
|
||||
|
||||
for i = 1, MAX_UPSTREAM_CHECKS_COUNT do
|
||||
local key = string.format("%s.%s.%s", ngx.now() + i, ngx.worker.pid(), math.random(999999))
|
||||
|
||||
local new_upstream = self.instance:find(key)
|
||||
|
||||
if not failed_upstreams[new_upstream] then
|
||||
return new_upstream, key
|
||||
end
|
||||
end
|
||||
|
||||
return nil, nil
|
||||
end
|
||||
|
||||
local function should_set_cookie(self)
|
||||
if self.cookie_session_affinity.locations and ngx.var.host then
|
||||
local locs = self.cookie_session_affinity.locations[ngx.var.host]
|
||||
|
|
@ -128,15 +107,9 @@ local function should_set_cookie(self)
|
|||
end
|
||||
|
||||
function _M.balance(self)
|
||||
local cookie, err = ck:new()
|
||||
if not cookie then
|
||||
ngx.log(ngx.ERR, "error while initializing cookie: " .. tostring(err))
|
||||
return
|
||||
end
|
||||
|
||||
local upstream_from_cookie
|
||||
|
||||
local key = cookie:get(self:cookie_name())
|
||||
local key = self:get_cookie()
|
||||
if key then
|
||||
upstream_from_cookie = self.instance:find(key)
|
||||
end
|
||||
|
|
@ -151,30 +124,22 @@ function _M.balance(self)
|
|||
|
||||
local new_upstream
|
||||
|
||||
new_upstream, key = pick_new_upstream(self)
|
||||
new_upstream, key = self:pick_new_upstream(get_failed_upstreams())
|
||||
if not new_upstream then
|
||||
ngx.log(ngx.WARN, string.format("failed to get new upstream; using upstream %s", new_upstream))
|
||||
elseif should_set_cookie(self) then
|
||||
set_cookie(self, key)
|
||||
self:set_cookie(key)
|
||||
end
|
||||
|
||||
return new_upstream
|
||||
end
|
||||
|
||||
function _M.sync(self, backend)
|
||||
-- reload balancer nodes
|
||||
balancer_resty.sync(self, backend)
|
||||
|
||||
-- Reload the balancer if any of the annotations have changed.
|
||||
local changed = not util.deep_compare(
|
||||
self.cookie_session_affinity,
|
||||
backend.sessionAffinityConfig.cookieSessionAffinity
|
||||
)
|
||||
if not changed then
|
||||
return
|
||||
end
|
||||
|
||||
ngx_log(INFO, string_format("[%s] nodes have changed for backend %s", self.name, backend.name))
|
||||
|
||||
self.traffic_shaping_policy = backend.trafficShapingPolicy
|
||||
self.alternative_backends = backend.alternativeBackends
|
||||
self.cookie_session_affinity = backend.sessionAffinityConfig.cookieSessionAffinity
|
||||
end
|
||||
|
||||
|
|
|
|||
49
rootfs/etc/nginx/lua/balancer/sticky_balanced.lua
Normal file
49
rootfs/etc/nginx/lua/balancer/sticky_balanced.lua
Normal file
|
|
@ -0,0 +1,49 @@
|
|||
-- An affinity mode which makes sure connections are rebalanced when a deployment is scaled.
|
||||
-- The advantage of this mode is that the load on the pods will be redistributed.
|
||||
-- The drawback of this mode is that, when scaling up a deployment, roughly (n-c)/n users
|
||||
-- will lose their session, where c is the current number of pods and n is the new number of
|
||||
-- pods.
|
||||
--
|
||||
local balancer_sticky = require("balancer.sticky")
|
||||
local math_random = require("math").random
|
||||
local resty_chash = require("resty.chash")
|
||||
local util_get_nodes = require("util").get_nodes
|
||||
|
||||
local _M = balancer_sticky:new()
|
||||
|
||||
-- Consider the situation of N upstreams one of which is failing.
|
||||
-- Then the probability to obtain failing upstream after M iterations would be close to (1/N)**M.
|
||||
-- For the worst case (2 upstreams; 20 iterations) it would be ~10**(-6)
|
||||
-- which is much better then ~10**(-3) for 10 iterations.
|
||||
local MAX_UPSTREAM_CHECKS_COUNT = 20
|
||||
|
||||
function _M.new(self, backend)
|
||||
local nodes = util_get_nodes(backend.endpoints)
|
||||
|
||||
local o = {
|
||||
name = "sticky_balanced",
|
||||
instance = resty_chash:new(nodes)
|
||||
}
|
||||
|
||||
setmetatable(o, self)
|
||||
self.__index = self
|
||||
|
||||
balancer_sticky.sync(o, backend)
|
||||
|
||||
return o
|
||||
end
|
||||
|
||||
function _M.pick_new_upstream(self, failed_upstreams)
|
||||
for i = 1, MAX_UPSTREAM_CHECKS_COUNT do
|
||||
local key = string.format("%s.%s.%s", ngx.now() + i, ngx.worker.pid(), math_random(999999))
|
||||
local new_upstream = self.instance:find(key)
|
||||
|
||||
if not failed_upstreams[new_upstream] then
|
||||
return new_upstream, key
|
||||
end
|
||||
end
|
||||
|
||||
return nil, nil
|
||||
end
|
||||
|
||||
return _M
|
||||
33
rootfs/etc/nginx/lua/balancer/sticky_persistent.lua
Normal file
33
rootfs/etc/nginx/lua/balancer/sticky_persistent.lua
Normal file
|
|
@ -0,0 +1,33 @@
|
|||
-- An affinity mode which makes sure a session is always routed to the same endpoint.
|
||||
-- The advantage of this mode is that a user will never lose his session.
|
||||
-- The drawback of this mode is that when scaling up a deployment, sessions will not
|
||||
-- be rebalanced.
|
||||
--
|
||||
local balancer_sticky = require("balancer.sticky")
|
||||
local util_get_nodes = require("util").get_nodes
|
||||
local util_nodemap = require("util.nodemap")
|
||||
|
||||
local _M = balancer_sticky:new()
|
||||
|
||||
function _M.new(self, backend)
|
||||
local nodes = util_get_nodes(backend.endpoints)
|
||||
local hash_salt = backend["name"]
|
||||
|
||||
local o = {
|
||||
name = "sticky_persistent",
|
||||
instance = util_nodemap:new(nodes, hash_salt)
|
||||
}
|
||||
|
||||
setmetatable(o, self)
|
||||
self.__index = self
|
||||
|
||||
balancer_sticky.sync(o, backend)
|
||||
|
||||
return o
|
||||
end
|
||||
|
||||
function _M.pick_new_upstream(self, failed_upstreams)
|
||||
return self.instance:random_except(failed_upstreams)
|
||||
end
|
||||
|
||||
return _M
|
||||
|
|
@ -1,4 +1,5 @@
|
|||
local sticky = require("balancer.sticky")
|
||||
local sticky_balanced = require("balancer.sticky_balanced")
|
||||
local sticky_persistent = require("balancer.sticky_persistent")
|
||||
local cookie = require("resty.cookie")
|
||||
local util = require("util")
|
||||
|
||||
|
|
@ -15,11 +16,16 @@ local function reset_ngx()
|
|||
end
|
||||
|
||||
function get_mocked_cookie_new()
|
||||
local o = { value = nil }
|
||||
local mock = {
|
||||
get = function(self, n) return self.value end,
|
||||
set = function(self, c) self.value = c.value ; return true, nil end
|
||||
}
|
||||
setmetatable(o, mock)
|
||||
mock.__index = mock
|
||||
|
||||
return function(self)
|
||||
return {
|
||||
get = function(self, n) return nil, "error" end,
|
||||
set = function(self, n) return true, "" end
|
||||
}
|
||||
return o;
|
||||
end
|
||||
end
|
||||
|
||||
|
|
@ -52,21 +58,27 @@ describe("Sticky", function()
|
|||
|
||||
describe("new(backend)", function()
|
||||
context("when backend specifies cookie name", function()
|
||||
it("returns an instance containing the corresponding cookie name", function()
|
||||
local function test(sticky)
|
||||
local sticky_balancer_instance = sticky:new(test_backend)
|
||||
local test_backend_cookie_name = test_backend.sessionAffinityConfig.cookieSessionAffinity.name
|
||||
assert.equal(sticky_balancer_instance:cookie_name(), test_backend_cookie_name)
|
||||
end)
|
||||
end
|
||||
|
||||
it("returns an instance containing the corresponding cookie name", function() test(sticky_balanced) end)
|
||||
it("returns an instance containing the corresponding cookie name", function() test(sticky_persistent) end)
|
||||
end)
|
||||
|
||||
context("when backend does not specify cookie name", function()
|
||||
it("returns an instance with 'route' as cookie name", function()
|
||||
local function test(sticky)
|
||||
local temp_backend = util.deepcopy(test_backend)
|
||||
temp_backend.sessionAffinityConfig.cookieSessionAffinity.name = nil
|
||||
local sticky_balancer_instance = sticky:new(temp_backend)
|
||||
local default_cookie_name = "route"
|
||||
assert.equal(sticky_balancer_instance:cookie_name(), default_cookie_name)
|
||||
end)
|
||||
end
|
||||
|
||||
it("returns an instance with 'route' as cookie name", function() test(sticky_balanced) end)
|
||||
it("returns an instance with 'route' as cookie name", function() test(sticky_persistent) end)
|
||||
end)
|
||||
end)
|
||||
|
||||
|
|
@ -74,8 +86,10 @@ describe("Sticky", function()
|
|||
local mocked_cookie_new = cookie.new
|
||||
|
||||
before_each(function()
|
||||
package.loaded["balancer.sticky"] = nil
|
||||
sticky = require("balancer.sticky")
|
||||
package.loaded["balancer.sticky_balanced"] = nil
|
||||
package.loaded["balancer.sticky_persistent"] = nil
|
||||
sticky_balanced = require("balancer.sticky_balanced")
|
||||
sticky_persistent = require("balancer.sticky_persistent")
|
||||
end)
|
||||
|
||||
after_each(function()
|
||||
|
|
@ -83,13 +97,17 @@ describe("Sticky", function()
|
|||
end)
|
||||
|
||||
context("when client doesn't have a cookie set and location is in cookie_locations", function()
|
||||
it("picks an endpoint for the client", function()
|
||||
|
||||
local function test_pick_endpoint(sticky)
|
||||
local sticky_balancer_instance = sticky:new(test_backend)
|
||||
local peer = sticky_balancer_instance:balance()
|
||||
assert.equal(peer, test_backend_endpoint)
|
||||
end)
|
||||
assert.equal(test_backend_endpoint, peer)
|
||||
end
|
||||
|
||||
it("sets a cookie on the client", function()
|
||||
it("picks an endpoint for the client", function() test_pick_endpoint(sticky_balanced) end)
|
||||
it("picks an endpoint for the client", function() test_pick_endpoint(sticky_persistent) end)
|
||||
|
||||
local function test_set_cookie(sticky)
|
||||
local s = {}
|
||||
cookie.new = function(self)
|
||||
local cookie_instance = {
|
||||
|
|
@ -112,9 +130,12 @@ describe("Sticky", function()
|
|||
local sticky_balancer_instance = sticky:new(b)
|
||||
assert.has_no.errors(function() sticky_balancer_instance:balance() end)
|
||||
assert.spy(s).was_called()
|
||||
end)
|
||||
end
|
||||
|
||||
it("sets a secure cookie on the client when being in ssl mode", function()
|
||||
it("sets a cookie on the client", function() test_set_cookie(sticky_balanced) end)
|
||||
it("sets a cookie on the client", function() test_set_cookie(sticky_persistent) end)
|
||||
|
||||
local function test_set_ssl_cookie(sticky)
|
||||
ngx.var.https = "on"
|
||||
local s = {}
|
||||
cookie.new = function(self)
|
||||
|
|
@ -138,10 +159,18 @@ describe("Sticky", function()
|
|||
local sticky_balancer_instance = sticky:new(b)
|
||||
assert.has_no.errors(function() sticky_balancer_instance:balance() end)
|
||||
assert.spy(s).was_called()
|
||||
end
|
||||
|
||||
it("sets a secure cookie on the client when being in ssl mode", function()
|
||||
test_set_ssl_cookie(sticky_balanced)
|
||||
end)
|
||||
it("sets a secure cookie on the client when being in ssl mode", function()
|
||||
test_set_ssl_cookie(sticky_persistent)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when client doesn't have a cookie set and cookie_locations contains a matching wildcard location", function()
|
||||
context("when client doesn't have a cookie set and cookie_locations contains a matching wildcard location",
|
||||
function()
|
||||
before_each(function ()
|
||||
ngx.var.host = "dev.test.com"
|
||||
end)
|
||||
|
|
@ -149,7 +178,7 @@ describe("Sticky", function()
|
|||
ngx.var.host = "test.com"
|
||||
end)
|
||||
|
||||
it("sets a cookie on the client", function()
|
||||
local function test(sticky)
|
||||
local s = {}
|
||||
cookie.new = function(self)
|
||||
local cookie_instance = {
|
||||
|
|
@ -173,17 +202,24 @@ describe("Sticky", function()
|
|||
local sticky_balancer_instance = sticky:new(b)
|
||||
assert.has_no.errors(function() sticky_balancer_instance:balance() end)
|
||||
assert.spy(s).was_called()
|
||||
end)
|
||||
end
|
||||
|
||||
it("sets a cookie on the client", function() test(sticky_balanced) end)
|
||||
it("sets a cookie on the client", function() test(sticky_persistent) end)
|
||||
end)
|
||||
|
||||
context("when client doesn't have a cookie set and location not in cookie_locations", function()
|
||||
it("picks an endpoint for the client", function()
|
||||
|
||||
local function test_pick_endpoint(sticky)
|
||||
local sticky_balancer_instance = sticky:new(test_backend)
|
||||
local peer = sticky_balancer_instance:balance()
|
||||
assert.equal(peer, test_backend_endpoint)
|
||||
end)
|
||||
end
|
||||
|
||||
it("does not set a cookie on the client", function()
|
||||
it("picks an endpoint for the client", function() test_pick_endpoint(sticky_balanced) end)
|
||||
it("picks an endpoint for the client", function() test_pick_endpoint(sticky_persistent) end)
|
||||
|
||||
local function test_no_cookie(sticky)
|
||||
local s = {}
|
||||
cookie.new = function(self)
|
||||
local cookie_instance = {
|
||||
|
|
@ -202,11 +238,15 @@ describe("Sticky", function()
|
|||
local sticky_balancer_instance = sticky:new(get_test_backend())
|
||||
assert.has_no.errors(function() sticky_balancer_instance:balance() end)
|
||||
assert.spy(s).was_not_called()
|
||||
end)
|
||||
end
|
||||
|
||||
it("does not set a cookie on the client", function() test_no_cookie(sticky_balanced) end)
|
||||
it("does not set a cookie on the client", function() test_no_cookie(sticky_persistent) end)
|
||||
end)
|
||||
|
||||
context("when client has a cookie set", function()
|
||||
it("does not set a cookie", function()
|
||||
|
||||
local function test_no_cookie(sticky)
|
||||
local s = {}
|
||||
cookie.new = function(self)
|
||||
local return_obj = {
|
||||
|
|
@ -219,13 +259,19 @@ describe("Sticky", function()
|
|||
local sticky_balancer_instance = sticky:new(test_backend)
|
||||
assert.has_no.errors(function() sticky_balancer_instance:balance() end)
|
||||
assert.spy(s).was_not_called()
|
||||
end)
|
||||
end
|
||||
|
||||
it("returns the correct endpoint for the client", function()
|
||||
it("does not set a cookie", function() test_no_cookie(sticky_balanced) end)
|
||||
it("does not set a cookie", function() test_no_cookie(sticky_persistent) end)
|
||||
|
||||
local function test_correct_endpoint(sticky)
|
||||
local sticky_balancer_instance = sticky:new(test_backend)
|
||||
local peer = sticky_balancer_instance:balance()
|
||||
assert.equal(peer, test_backend_endpoint)
|
||||
end)
|
||||
end
|
||||
|
||||
it("returns the correct endpoint for the client", function() test_correct_endpoint(sticky_balanced) end)
|
||||
it("returns the correct endpoint for the client", function() test_correct_endpoint(sticky_persistent) end)
|
||||
end)
|
||||
end)
|
||||
|
||||
|
|
@ -238,7 +284,12 @@ describe("Sticky", function()
|
|||
},
|
||||
sessionAffinityConfig = {
|
||||
name = "cookie",
|
||||
cookieSessionAffinity = { name = "test_name", hash = "sha1", change_on_failure = change_on_failure }
|
||||
cookieSessionAffinity = {
|
||||
name = "test_name",
|
||||
hash = "sha1",
|
||||
change_on_failure = change_on_failure,
|
||||
locations = { ['test.com'] = {'/'} }
|
||||
}
|
||||
},
|
||||
}
|
||||
end
|
||||
|
|
@ -247,53 +298,58 @@ describe("Sticky", function()
|
|||
local mocked_cookie_new = cookie.new
|
||||
|
||||
before_each(function()
|
||||
package.loaded["balancer.sticky"] = nil
|
||||
sticky = require("balancer.sticky")
|
||||
package.loaded["balancer.sticky_balanced"] = nil
|
||||
package.loaded["balancer.sticky_persistent"] = nil
|
||||
sticky_balanced = require("balancer.sticky_balanced")
|
||||
sticky_persistent = require("balancer.sticky_persistent")
|
||||
mock_ngx({ var = { location_path = "/", host = "test.com" } })
|
||||
end)
|
||||
|
||||
after_each(function()
|
||||
cookie.new = mocked_cookie_new
|
||||
reset_ngx()
|
||||
end)
|
||||
|
||||
context("when request to upstream fails", function()
|
||||
|
||||
local function test(sticky, change_on_failure)
|
||||
local sticky_balancer_instance = sticky:new(get_several_test_backends(change_on_failure))
|
||||
|
||||
local old_upstream = sticky_balancer_instance:balance()
|
||||
assert.is.Not.Nil(old_upstream)
|
||||
for _ = 1, 100 do
|
||||
-- make sure upstream doesn't change on subsequent calls of balance()
|
||||
assert.equal(old_upstream, sticky_balancer_instance:balance())
|
||||
end
|
||||
|
||||
-- simulate request failure
|
||||
sticky_balancer_instance.get_last_failure = function()
|
||||
return "failed"
|
||||
end
|
||||
_G.ngx.var.upstream_addr = old_upstream
|
||||
|
||||
for _ = 1, 100 do
|
||||
local new_upstream = sticky_balancer_instance:balance()
|
||||
if change_on_failure == false then
|
||||
-- upstream should be the same inspite of error, if change_on_failure option is false
|
||||
assert.equal(new_upstream, old_upstream)
|
||||
else
|
||||
-- upstream should change after error, if change_on_failure option is true
|
||||
assert.not_equal(new_upstream, old_upstream)
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
it("changes upstream when change_on_failure option is true", function()
|
||||
-- create sticky cookie
|
||||
cookie.new = function(self)
|
||||
local return_obj = {
|
||||
set = function(v) return false, nil end,
|
||||
get = function(k) return "" end,
|
||||
}
|
||||
return return_obj, false
|
||||
end
|
||||
|
||||
local options = {false, true}
|
||||
|
||||
for _, option in ipairs(options) do
|
||||
local sticky_balancer_instance = sticky:new(get_several_test_backends(option))
|
||||
|
||||
local old_upstream = sticky_balancer_instance:balance()
|
||||
for _ = 1, 100 do
|
||||
-- make sure upstream doesn't change on subsequent calls of balance()
|
||||
assert.equal(old_upstream, sticky_balancer_instance:balance())
|
||||
end
|
||||
|
||||
-- simulate request failure
|
||||
sticky_balancer_instance.get_last_failure = function()
|
||||
return "failed"
|
||||
end
|
||||
_G.ngx.var = { upstream_addr = old_upstream }
|
||||
|
||||
for _ = 1, 100 do
|
||||
local new_upstream = sticky_balancer_instance:balance()
|
||||
if option == false then
|
||||
-- upstream should be the same inspite of error, if change_on_failure option is false
|
||||
assert.equal(new_upstream, old_upstream)
|
||||
else
|
||||
-- upstream should change after error, if change_on_failure option is true
|
||||
assert.not_equal(new_upstream, old_upstream)
|
||||
end
|
||||
end
|
||||
end
|
||||
test(sticky_balanced, true)
|
||||
end)
|
||||
it("changes upstream when change_on_failure option is true", function()
|
||||
test(sticky_balanced, false)
|
||||
end)
|
||||
it("changes upstream when change_on_failure option is true", function()
|
||||
test(sticky_persistent, true)
|
||||
end)
|
||||
it("changes upstream when change_on_failure option is true", function()
|
||||
test(sticky_persistent, false)
|
||||
end)
|
||||
end)
|
||||
end)
|
||||
|
|
|
|||
|
|
@ -23,9 +23,9 @@ local function reset_expected_implementations()
|
|||
["access-router-production-web-80"] = package.loaded["balancer.round_robin"],
|
||||
["my-dummy-app-1"] = package.loaded["balancer.round_robin"],
|
||||
["my-dummy-app-2"] = package.loaded["balancer.chash"],
|
||||
["my-dummy-app-3"] = package.loaded["balancer.sticky"],
|
||||
["my-dummy-app-3"] = package.loaded["balancer.sticky_persistent"],
|
||||
["my-dummy-app-4"] = package.loaded["balancer.ewma"],
|
||||
["my-dummy-app-5"] = package.loaded["balancer.sticky"],
|
||||
["my-dummy-app-5"] = package.loaded["balancer.sticky_balanced"]
|
||||
}
|
||||
end
|
||||
|
||||
|
|
@ -55,7 +55,7 @@ local function reset_backends()
|
|||
},
|
||||
{
|
||||
name = "my-dummy-app-3", ["load-balance"] = "ewma",
|
||||
sessionAffinityConfig = { name = "cookie", cookieSessionAffinity = { name = "route" } }
|
||||
sessionAffinityConfig = { name = "cookie", mode = 'persistent', cookieSessionAffinity = { name = "route" } }
|
||||
},
|
||||
{ name = "my-dummy-app-4", ["load-balance"] = "ewma", },
|
||||
{
|
||||
|
|
|
|||
167
rootfs/etc/nginx/lua/test/util/nodemap_test.lua
Normal file
167
rootfs/etc/nginx/lua/test/util/nodemap_test.lua
Normal file
|
|
@ -0,0 +1,167 @@
|
|||
local util = require("util")
|
||||
local nodemap = require("util.nodemap")
|
||||
|
||||
local function get_test_backend_single()
|
||||
return {
|
||||
name = "access-router-production-web-80",
|
||||
endpoints = {
|
||||
{ address = "10.184.7.40", port = "8080", maxFails = 0, failTimeout = 0 }
|
||||
}
|
||||
}
|
||||
end
|
||||
|
||||
local function get_test_backend_multi()
|
||||
return {
|
||||
name = "access-router-production-web-80",
|
||||
endpoints = {
|
||||
{ address = "10.184.7.40", port = "8080", maxFails = 0, failTimeout = 0 },
|
||||
{ address = "10.184.7.41", port = "8080", maxFails = 0, failTimeout = 0 }
|
||||
}
|
||||
}
|
||||
end
|
||||
|
||||
local function get_test_nodes_ignore(endpoint)
|
||||
local ignore = {}
|
||||
ignore[endpoint] = true
|
||||
return ignore
|
||||
end
|
||||
|
||||
describe("Node Map", function()
|
||||
|
||||
local test_backend_single = get_test_backend_single()
|
||||
local test_backend_multi = get_test_backend_multi()
|
||||
local test_salt = test_backend_single.name
|
||||
local test_nodes_single = util.get_nodes(test_backend_single.endpoints)
|
||||
local test_nodes_multi = util.get_nodes(test_backend_multi.endpoints)
|
||||
local test_endpoint1 = test_backend_multi.endpoints[1].address .. ":" .. test_backend_multi.endpoints[1].port
|
||||
local test_endpoint2 = test_backend_multi.endpoints[2].address .. ":" .. test_backend_multi.endpoints[2].port
|
||||
local test_nodes_ignore = get_test_nodes_ignore(test_endpoint1)
|
||||
|
||||
describe("new()", function()
|
||||
context("when no salt has been provided", function()
|
||||
it("random() returns an unsalted key", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, nil)
|
||||
local expected_endpoint = test_endpoint1
|
||||
local expected_hash_key = ngx.md5(expected_endpoint)
|
||||
local actual_endpoint
|
||||
local actual_hash_key
|
||||
|
||||
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||
|
||||
assert.equal(actual_endpoint, expected_endpoint)
|
||||
assert.equal(expected_hash_key, actual_hash_key)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when a salt has been provided", function()
|
||||
it("random() returns a salted key", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||
local expected_endpoint = test_endpoint1
|
||||
local expected_hash_key = ngx.md5(test_salt .. expected_endpoint)
|
||||
local actual_endpoint
|
||||
local actual_hash_key
|
||||
|
||||
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||
|
||||
assert.equal(actual_endpoint, expected_endpoint)
|
||||
assert.equal(expected_hash_key, actual_hash_key)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when no nodes have been provided", function()
|
||||
it("random() returns nil", function()
|
||||
local nodemap_instance = nodemap:new({}, test_salt)
|
||||
local actual_endpoint
|
||||
local actual_hash_key
|
||||
|
||||
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||
|
||||
assert.equal(actual_endpoint, nil)
|
||||
assert.equal(expected_hash_key, nil)
|
||||
end)
|
||||
end)
|
||||
end)
|
||||
|
||||
describe("find()", function()
|
||||
before_each(function()
|
||||
package.loaded["util.nodemap"] = nil
|
||||
nodemap = require("util.nodemap")
|
||||
end)
|
||||
|
||||
context("when a hash key is valid", function()
|
||||
it("find() returns the correct endpoint", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||
local test_hash_key
|
||||
local expected_endpoint
|
||||
local actual_endpoint
|
||||
|
||||
expected_endpoint, test_hash_key = nodemap_instance:random()
|
||||
assert.not_equal(expected_endpoint, nil)
|
||||
assert.not_equal(test_hash_key, nil)
|
||||
|
||||
actual_endpoint = nodemap_instance:find(test_hash_key)
|
||||
assert.equal(actual_endpoint, expected_endpoint)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when a hash key is invalid", function()
|
||||
it("find() returns nil", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||
local test_hash_key = "invalid or nonexistent hash key"
|
||||
local actual_endpoint
|
||||
|
||||
actual_endpoint = nodemap_instance:find(test_hash_key)
|
||||
|
||||
assert.equal(actual_endpoint, nil)
|
||||
end)
|
||||
end)
|
||||
end)
|
||||
|
||||
|
||||
describe("random_except()", function()
|
||||
before_each(function()
|
||||
package.loaded["util.nodemap"] = nil
|
||||
nodemap = require("util.nodemap")
|
||||
end)
|
||||
|
||||
context("when nothing has been excluded", function()
|
||||
it("random_except() returns the correct endpoint", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||
local expected_endpoint = test_endpoint1
|
||||
local test_hash_key
|
||||
local actual_endpoint
|
||||
|
||||
actual_endpoint, test_hash_key = nodemap_instance:random_except({})
|
||||
assert.equal(expected_endpoint, actual_endpoint)
|
||||
assert.not_equal(test_hash_key, nil)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when everything has been excluded", function()
|
||||
it("random_except() returns nil", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||
local actual_hash_key
|
||||
local actual_endpoint
|
||||
|
||||
actual_endpoint, actual_hash_key = nodemap_instance:random_except(test_nodes_ignore)
|
||||
|
||||
assert.equal(actual_endpoint, nil)
|
||||
assert.equal(actual_hash_key, nil)
|
||||
end)
|
||||
end)
|
||||
|
||||
context("when an endpoint has been excluded", function()
|
||||
it("random_except() does not return it", function()
|
||||
local nodemap_instance = nodemap:new(test_nodes_multi, test_salt)
|
||||
local expected_endpoint = test_endpoint2
|
||||
local actual_endpoint
|
||||
local test_hash_key
|
||||
|
||||
actual_endpoint, test_hash_key = nodemap_instance:random_except(test_nodes_ignore)
|
||||
|
||||
assert.equal(actual_endpoint, expected_endpoint)
|
||||
assert.not_equal(test_hash_key, nil)
|
||||
end)
|
||||
end)
|
||||
end)
|
||||
end)
|
||||
119
rootfs/etc/nginx/lua/util/nodemap.lua
Normal file
119
rootfs/etc/nginx/lua/util/nodemap.lua
Normal file
|
|
@ -0,0 +1,119 @@
|
|||
local math_random = require("math").random
|
||||
local util_tablelength = require("util").tablelength
|
||||
|
||||
local _M = {}
|
||||
|
||||
--- create_map generates the node hash table
|
||||
-- @tparam {[string]=number} nodes A table with the node as a key and its weight as a value.
|
||||
-- @tparam string salt A salt that will be used to generate salted hash keys.
|
||||
local function create_map(nodes, salt)
|
||||
local hash_map = {}
|
||||
|
||||
for endpoint, _ in pairs(nodes) do
|
||||
-- obfuscate the endpoint with a shared key to prevent brute force
|
||||
-- and rainbow table attacks which could reveal internal endpoints
|
||||
local key = salt .. endpoint
|
||||
local hash_key = ngx.md5(key)
|
||||
hash_map[hash_key] = endpoint
|
||||
end
|
||||
|
||||
return hash_map
|
||||
end
|
||||
|
||||
--- get_random_node picks a random node from the given map.
|
||||
-- @tparam {[string], ...} map A key to node hash table.
|
||||
-- @treturn string,string The node and its key
|
||||
local function get_random_node(map)
|
||||
local size = util_tablelength(map)
|
||||
|
||||
if size < 1 then
|
||||
return nil, nil
|
||||
end
|
||||
|
||||
local index = math_random(1, size)
|
||||
local count = 1
|
||||
|
||||
for key, endpoint in pairs(map) do
|
||||
if count == index then
|
||||
return endpoint, key
|
||||
end
|
||||
|
||||
count = count + 1
|
||||
end
|
||||
|
||||
ngx.log(ngx.ERR, string.format("Failed to find node %d of %d! This is a bug, please report!", index, size))
|
||||
|
||||
return nil, nil
|
||||
end
|
||||
|
||||
--- new constructs a new instance of the node map
|
||||
--
|
||||
-- The map uses MD5 to create hash keys for a given node. For security reasons it supports
|
||||
-- salted hash keys, to prevent attackers from using rainbow tables or brute forcing
|
||||
-- the node endpoints, which would reveal cluster internal network information.
|
||||
--
|
||||
-- To make sure hash keys are reproducible on different ingress controller instances the salt
|
||||
-- needs to be shared and therefore is not simply generated randomly.
|
||||
--
|
||||
-- @tparam {[string]=number} endpoints A table with the node endpoint as a key and its weight as a value.
|
||||
-- @tparam[opt] string hash_salt A optional hash salt that will be used to obfuscate the hash key.
|
||||
function _M.new(self, endpoints, hash_salt)
|
||||
if hash_salt == nil then
|
||||
hash_salt = ''
|
||||
end
|
||||
|
||||
-- the endpoints have to be saved as 'nodes' to keep compatibility to balancer.resty
|
||||
local o = {
|
||||
salt = hash_salt,
|
||||
nodes = endpoints,
|
||||
map = create_map(endpoints, hash_salt)
|
||||
}
|
||||
|
||||
setmetatable(o, self)
|
||||
self.__index = self
|
||||
return o
|
||||
end
|
||||
|
||||
--- reinit reinitializes the node map reusing the original salt
|
||||
-- @tparam {[string]=number} nodes A table with the node as a key and its weight as a value.
|
||||
function _M.reinit(self, nodes)
|
||||
self.nodes = nodes
|
||||
self.map = create_map(nodes, self.salt)
|
||||
end
|
||||
|
||||
--- find looks up a node by hash key.
|
||||
-- @tparam string key The hash key.
|
||||
-- @treturn string The node.
|
||||
function _M.find(self, key)
|
||||
return self.map[key]
|
||||
end
|
||||
|
||||
--- random picks a random node from the hashmap.
|
||||
-- @treturn string,string A random node and its key or both nil.
|
||||
function _M.random(self)
|
||||
return get_random_node(self.map)
|
||||
end
|
||||
|
||||
--- random_except picks a random node from the hashmap, ignoring the nodes in the given table
|
||||
-- @tparam {string, } ignore_nodes A table of nodes to ignore, the node needs to be the key,
|
||||
-- the value needs to be set to true
|
||||
-- @treturn string,string A random node and its key or both nil.
|
||||
function _M.random_except(self, ignore_nodes)
|
||||
local valid_nodes = {}
|
||||
|
||||
-- avoid generating the map if no ignores where provided
|
||||
if ignore_nodes == nil or util_tablelength(ignore_nodes) == 0 then
|
||||
return get_random_node(self.map)
|
||||
end
|
||||
|
||||
-- generate valid endpoints
|
||||
for key, endpoint in pairs(self.map) do
|
||||
if not ignore_nodes[endpoint] then
|
||||
valid_nodes[key] = endpoint
|
||||
end
|
||||
end
|
||||
|
||||
return get_random_node(valid_nodes)
|
||||
end
|
||||
|
||||
return _M
|
||||
Loading…
Add table
Add a link
Reference in a new issue