Skip to content

Commit 0690f2b

Browse files
committed
Refactor litellm_ai_handler.py and update requirements.txt
- Replace retry library with tenacity for better exception handling - Add verbosity level checks for logging prompts and AI responses - Add support for HuggingFace API base and repetition penalty in chat completion - Update requirements.txt with tenacity library
1 parent 26fb2a4 commit 0690f2b

File tree

2 files changed

+28
-12
lines changed

2 files changed

+28
-12
lines changed

pr_agent/algo/ai_handlers/litellm_ai_handler.py

+27-12
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@
44
import litellm
55
import openai
66
from litellm import acompletion
7-
# from openai.error import APIError, RateLimitError, Timeout, TryAgain
8-
from openai import APIError, RateLimitError, Timeout
9-
from retry import retry
7+
from tenacity import retry, retry_if_exception_type, stop_after_attempt
108
from pr_agent.algo.ai_handlers.base_ai_handler import BaseAiHandler
119
from pr_agent.config_loader import get_settings
1210
from pr_agent.log import get_logger
@@ -28,7 +26,8 @@ def __init__(self):
2826
"""
2927
self.azure = False
3028
self.aws_bedrock_client = None
31-
29+
self.api_base = None
30+
self.repetition_penalty = None
3231
if get_settings().get("OPENAI.KEY", None):
3332
openai.api_key = get_settings().openai.key
3433
litellm.openai_key = get_settings().openai.key
@@ -57,8 +56,11 @@ def __init__(self):
5756
litellm.replicate_key = get_settings().replicate.key
5857
if get_settings().get("HUGGINGFACE.KEY", None):
5958
litellm.huggingface_key = get_settings().huggingface.key
60-
if get_settings().get("HUGGINGFACE.API_BASE", None):
61-
litellm.api_base = get_settings().huggingface.api_base
59+
if get_settings().get("HUGGINGFACE.API_BASE", None) and 'huggingface' in get_settings().config.model:
60+
litellm.api_base = get_settings().huggingface.api_base
61+
self.api_base = get_settings().huggingface.api_base
62+
if get_settings().get("HUGGINGFACE.REPITITION_PENALTY", None):
63+
self.repetition_penalty = float(get_settings().huggingface.repetition_penalty)
6264
if get_settings().get("VERTEXAI.VERTEX_PROJECT", None):
6365
litellm.vertex_project = get_settings().vertexai.vertex_project
6466
litellm.vertex_location = get_settings().get(
@@ -78,8 +80,10 @@ def deployment_id(self):
7880
"""
7981
return get_settings().get("OPENAI.DEPLOYMENT_ID", None)
8082

81-
@retry(exceptions=(APIError, Timeout, AttributeError, RateLimitError),
82-
tries=OPENAI_RETRIES, delay=2, backoff=2, jitter=(1, 3))
83+
@retry(
84+
retry=retry_if_exception_type((openai.APIError, openai.APIConnectionError, openai.Timeout)), # No retry on RateLimitError
85+
stop=stop_after_attempt(OPENAI_RETRIES)
86+
)
8387
async def chat_completion(self, model: str, system: str, user: str, temperature: float = 0.2):
8488
try:
8589
resp, finish_reason = None, None
@@ -93,28 +97,39 @@ async def chat_completion(self, model: str, system: str, user: str, temperature:
9397
"messages": messages,
9498
"temperature": temperature,
9599
"force_timeout": get_settings().config.ai_timeout,
100+
"api_base" : self.api_base,
96101
}
97102
if self.aws_bedrock_client:
98103
kwargs["aws_bedrock_client"] = self.aws_bedrock_client
104+
if self.repetition_penalty:
105+
kwargs["repetition_penalty"] = self.repetition_penalty
99106

100107
get_logger().debug("Prompts", artifact={"system": system, "user": user})
108+
109+
if get_settings().config.verbosity_level >= 2:
110+
get_logger().info(f"\nSystem prompt:\n{system}")
111+
get_logger().info(f"\nUser prompt:\n{user}")
112+
101113
response = await acompletion(**kwargs)
102-
except (APIError, Timeout) as e:
114+
except (openai.APIError, openai.Timeout) as e:
103115
get_logger().error("Error during OpenAI inference: ", e)
104116
raise
105-
except (RateLimitError) as e:
117+
except (openai.RateLimitError) as e:
106118
get_logger().error("Rate limit error during OpenAI inference: ", e)
107119
raise
108120
except (Exception) as e:
109121
get_logger().error("Unknown error during OpenAI inference: ", e)
110-
raise APIError from e
122+
raise openai.APIError from e
111123
if response is None or len(response["choices"]) == 0:
112-
raise APIError
124+
raise openai.APIError
113125
else:
114126
resp = response["choices"][0]['message']['content']
115127
finish_reason = response["choices"][0]["finish_reason"]
116128
# usage = response.get("usage")
117129
get_logger().debug(f"\nAI response:\n{resp}")
118130
get_logger().debug("Full_response", artifact=response)
119131

132+
if get_settings().config.verbosity_level >= 2:
133+
get_logger().info(f"\nAI response:\n{resp}")
134+
120135
return resp, finish_reason

requirements.txt

+1
Original file line numberDiff line numberDiff line change
@@ -25,4 +25,5 @@ starlette-context==0.3.6
2525
tiktoken==0.5.2
2626
ujson==5.8.0
2727
uvicorn==0.22.0
28+
tenacity==8.2.3
2829
# langchain==0.0.349 # uncomment this to support language LangChainOpenAIHandler

0 commit comments

Comments
 (0)