From 391388e7dc800fa1fedf4f155ba8fcf60d753dba Mon Sep 17 00:00:00 2001 From: Sicheng Song Date: Tue, 14 Nov 2023 23:12:58 +0000 Subject: [PATCH] Change rate limiter token capacity setting Signed-off-by: Sicheng Song --- .../main/java/org/opensearch/ml/model/MLModelManager.java | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/plugin/src/main/java/org/opensearch/ml/model/MLModelManager.java b/plugin/src/main/java/org/opensearch/ml/model/MLModelManager.java index be3eba6c6e..75c9a90c0e 100644 --- a/plugin/src/main/java/org/opensearch/ml/model/MLModelManager.java +++ b/plugin/src/main/java/org/opensearch/ml/model/MLModelManager.java @@ -1124,12 +1124,13 @@ private TokenBucket rateLimiterConstructor(Integer eligibleNodeCount, MLModel ml TimeUnit rateLimitUnit = mlModel.getRateLimitUnit(); log .debug( - "Initializing the rate limiter for Model {}, with TPS limit {}, evenly distributed on {} nodes", + "Initializing the rate limiter for Model {}, with TPS limit {} and burst capacity {}, evenly distributed on {} nodes", mlModel.getModelId(), rateLimitNumber / rateLimitUnit.toSeconds(1), + rateLimitNumber, eligibleNodeCount ); - return new TokenBucket(System::nanoTime, rateLimitNumber / rateLimitUnit.toNanos(1) / eligibleNodeCount, 2); + return new TokenBucket(System::nanoTime, rateLimitNumber / rateLimitUnit.toNanos(1) / eligibleNodeCount, rateLimitNumber); } return null; }