Exception Mapping
LiteLLM maps exceptions across all providers to their OpenAI counterparts.
All exceptions can be imported from litellm
- e.g. from litellm import BadRequestError
LiteLLM Exceptionsโ
Status Code | Error Type | Inherits from | Description |
---|---|---|---|
400 | BadRequestError | openai.BadRequestError | |
400 | UnsupportedParamsError | litellm.BadRequestError | Raised when unsupported params are passed |
400 | ContextWindowExceededError | litellm.BadRequestError | Special error type for context window exceeded error messages - enables context window fallbacks |
400 | ContentPolicyViolationError | litellm.BadRequestError | Special error type for content policy violation error messages - enables content policy fallbacks |
400 | InvalidRequestError | openai.BadRequestError | Deprecated error, use BadRequestError instead |
401 | AuthenticationError | openai.AuthenticationError | |
403 | PermissionDeniedError | openai.PermissionDeniedError | |
404 | NotFoundError | openai.NotFoundError | raise when invalid models passed, example gpt-8 |
408 | Timeout | openai.APITimeoutError | Raised when a timeout occurs |
422 | UnprocessableEntityError | openai.UnprocessableEntityError | |
429 | RateLimitError | openai.RateLimitError | |
500 | APIConnectionError | openai.APIConnectionError | If any unmapped error is returned, we return this error |
500 | APIError | openai.APIError | Generic 500-status code error |
503 | ServiceUnavailableError | openai.APIStatusError | If provider returns a service unavailable error, this error is raised |
>=500 | InternalServerError | openai.InternalServerError | If any unmapped 500-status code error is returned, this error is raised |
N/A | APIResponseValidationError | openai.APIResponseValidationError | If Rules are used, and request/response fails a rule, this error is raised |
N/A | BudgetExceededError | Exception | Raised for proxy, when budget is exceeded |
N/A | JSONSchemaValidationError | litellm.APIResponseValidationError | Raised when response does not match expected json schema - used if response_schema param passed in with enforce_validation=True |
N/A | MockException | Exception | Internal exception, raised by mock_completion class. Do not use directly |
N/A | OpenAIError | openai.OpenAIError | Deprecated internal exception, inherits from openai.OpenAIError. |
Base case we return APIConnectionError
All our exceptions inherit from OpenAI's exception types, so any error-handling you have for that, should work out of the box with LiteLLM.
For all cases, the exception returned inherits from the original OpenAI Exception but contains 3 additional attributes:
- status_code - the http status code of the exception
- message - the error message
- llm_provider - the provider raising the exception
Usageโ
import litellm
import openai
try:
response = litellm.completion(
model="gpt-4",
messages=[
{
"role": "user",
"content": "hello, write a 20 pageg essay"
}
],
timeout=0.01, # this will raise a timeout exception
)
except openai.APITimeoutError as e:
print("Passed: Raised correct exception. Got openai.APITimeoutError\nGood Job", e)
print(type(e))
pass
Usage - Catching Streaming Exceptionsโ
import litellm
try:
response = litellm.completion(
model="gpt-3.5-turbo",
messages=[
{
"role": "user",
"content": "hello, write a 20 pg essay"
}
],
timeout=0.0001, # this will raise an exception
stream=True,
)
for chunk in response:
print(chunk)
except openai.APITimeoutError as e:
print("Passed: Raised correct exception. Got openai.APITimeoutError\nGood Job", e)
print(type(e))
pass
except Exception as e:
print(f"Did not raise error `openai.APITimeoutError`. Instead raised error type: {type(e)}, Error: {e}")
Usage - Should you retry exception?โ
import litellm
import openai
try:
response = litellm.completion(
model="gpt-4",
messages=[
{
"role": "user",
"content": "hello, write a 20 pageg essay"
}
],
timeout=0.01, # this will raise a timeout exception
)
except openai.APITimeoutError as e:
should_retry = litellm._should_retry(e.status_code)
print(f"should_retry: {should_retry}")
Advancedโ
Accessing Provider-Specific Error Detailsโ
LiteLLM exceptions include a provider_specific_fields
attribute that contains additional error information specific to each provider. This is particularly useful for Azure OpenAI, which provides detailed content filtering information.
Azure OpenAI - Content Policy Violation Inner Error Accessโ
When Azure OpenAI returns content policy violations, you can access the detailed content filtering results through the innererror
field:
import litellm
from litellm.exceptions import ContentPolicyViolationError
try:
response = litellm.completion(
model="azure/gpt-4",
messages=[
{
"role": "user",
"content": "Some content that might violate policies"
}
]
)
except ContentPolicyViolationError as e:
# Access Azure-specific error details
if e.provider_specific_fields and "innererror" in e.provider_specific_fields:
innererror = e.provider_specific_fields["innererror"]
# Access content filter results
content_filter_result = innererror.get("content_filter_result", {})
print(f"Content filter code: {innererror.get('code')}")
print(f"Hate filtered: {content_filter_result.get('hate', {}).get('filtered')}")
print(f"Violence severity: {content_filter_result.get('violence', {}).get('severity')}")
print(f"Sexual content filtered: {content_filter_result.get('sexual', {}).get('filtered')}")
Example Response Structure:
When calling the LiteLLM proxy, content policy violations will return detailed filtering information:
{
"error": {
"message": "litellm.ContentPolicyViolationError: AzureException - The response was filtered due to the prompt triggering Azure OpenAI's content management policy...",
"type": null,
"param": null,
"code": "400",
"provider_specific_fields": {
"innererror": {
"code": "ResponsibleAIPolicyViolation",
"content_filter_result": {
"hate": {
"filtered": true,
"severity": "high"
},
"jailbreak": {
"filtered": false,
"detected": false
},
"self_harm": {
"filtered": false,
"severity": "safe"
},
"sexual": {
"filtered": false,
"severity": "safe"
},
"violence": {
"filtered": true,
"severity": "medium"
}
}
}
}
}
}
## Details
To see how it's implemented - [check out the code](https://github.com/BerriAI/litellm/blob/a42c197e5a6de56ea576c73715e6c7c6b19fa249/litellm/utils.py#L1217)
[Create an issue](https://github.com/BerriAI/litellm/issues/new) **or** [make a PR](https://github.com/BerriAI/litellm/pulls) if you want to improve the exception mapping.
**Note** For OpenAI and Azure we return the original exception (since they're of the OpenAI Error type). But we add the 'llm_provider' attribute to them. [See code](https://github.com/BerriAI/litellm/blob/a42c197e5a6de56ea576c73715e6c7c6b19fa249/litellm/utils.py#L1221)
## Custom mapping list
Base case - we return `litellm.APIConnectionError` exception (inherits from openai's APIConnectionError exception).
| custom_llm_provider | Timeout | ContextWindowExceededError | BadRequestError | NotFoundError | ContentPolicyViolationError | AuthenticationError | APIError | RateLimitError | ServiceUnavailableError | PermissionDeniedError | UnprocessableEntityError |
|----------------------------|---------|----------------------------|------------------|---------------|-----------------------------|---------------------|----------|----------------|-------------------------|-----------------------|-------------------------|
| openai | โ | โ | โ | | โ | โ | | | | | |
| watsonx | | | | | | | |โ| | | |
| text-completion-openai | โ | โ | โ | | โ | โ | | | | | |
| custom_openai | โ | โ | โ | | โ | โ | | | | | |
| openai_compatible_providers| โ | โ | โ | | โ | โ | | | | | |
| anthropic | โ | โ | โ | โ | | โ | | | โ | โ | |
| replicate | โ | โ | โ | โ | | โ | | โ | โ | | |
| bedrock | โ | โ | โ | โ | | โ | | โ | โ | โ | |
| sagemaker | | โ | โ | | | | | | | | |
| vertex_ai | โ | | โ | | | | โ | | | | โ |
| palm | โ | โ | | | | | โ | | | | |
| gemini | โ | โ | | | | | โ | | | | |
| cloudflare | | | โ | | | โ | | | | | |
| cohere | | โ | โ | | | โ | | | โ | | |
| cohere_chat | | โ | โ | | | โ | | | โ | | |
| huggingface | โ | โ | โ | | | โ | | โ | โ | | |
| ai21 | โ | โ | โ | โ | | โ | | โ | | | |
| nlp_cloud | โ | โ | โ | | | โ | โ | โ | โ | | |
| together_ai | โ | โ | โ | | | โ | | | | | |
| aleph_alpha | | | โ | | | โ | | | | | |
| ollama | โ | | โ | | | | | | โ | | |
| ollama_chat | โ | | โ | | | | | | โ | | |
| vllm | | | | | | โ | โ | | | | |
| azure | โ | โ | โ | โ | โ | โ | | | โ | | |
- "โ" indicates that the specified `custom_llm_provider` can raise the corresponding exception.
- Empty cells indicate the lack of association or that the provider does not raise that particular exception type as indicated by the function.
> For a deeper understanding of these exceptions, you can check out [this](https://github.com/BerriAI/litellm/blob/d7e58d13bf9ba9edbab2ab2f096f3de7547f35fa/litellm/utils.py#L1544) implementation for additional insights.
The `ContextWindowExceededError` is a sub-class of `InvalidRequestError`. It was introduced to provide more granularity for exception-handling scenarios. Please refer to [this issue to learn more](https://github.com/BerriAI/litellm/issues/228).
Contributions to improve exception mapping are [welcome](https://github.com/BerriAI/litellm#contributing)