Spaces:
Configuration error
Configuration error
import io | |
import os | |
import sys | |
from typing import Optional | |
sys.path.insert(0, os.path.abspath("../..")) | |
import asyncio | |
import gzip | |
import json | |
import logging | |
import time | |
from unittest.mock import AsyncMock, patch | |
import pytest | |
import litellm | |
from litellm._logging import verbose_logger | |
from litellm.integrations.custom_logger import CustomLogger | |
from litellm.types.utils import StandardLoggingPayload | |
class TestCustomLogger(CustomLogger): | |
def __init__(self, *args, **kwargs): | |
super().__init__(*args, **kwargs) | |
self.logged_standard_logging_payload: Optional[StandardLoggingPayload] = None | |
async def async_log_success_event(self, kwargs, response_obj, start_time, end_time): | |
standard_logging_payload = kwargs.get("standard_logging_object", None) | |
self.logged_standard_logging_payload = standard_logging_payload | |
async def test_global_redaction_on(): | |
litellm.turn_off_message_logging = True | |
test_custom_logger = TestCustomLogger() | |
litellm.callbacks = [test_custom_logger] | |
response = await litellm.acompletion( | |
model="gpt-3.5-turbo", | |
messages=[{"role": "user", "content": "hi"}], | |
mock_response="hello", | |
) | |
await asyncio.sleep(1) | |
standard_logging_payload = test_custom_logger.logged_standard_logging_payload | |
assert standard_logging_payload is not None | |
assert standard_logging_payload["response"] == {"text": "redacted-by-litellm"} | |
assert standard_logging_payload["messages"][0]["content"] == "redacted-by-litellm" | |
print( | |
"logged standard logging payload", | |
json.dumps(standard_logging_payload, indent=2), | |
) | |
async def test_global_redaction_with_dynamic_params(turn_off_message_logging): | |
litellm.turn_off_message_logging = True | |
test_custom_logger = TestCustomLogger() | |
litellm.callbacks = [test_custom_logger] | |
response = await litellm.acompletion( | |
model="gpt-3.5-turbo", | |
messages=[{"role": "user", "content": "hi"}], | |
turn_off_message_logging=turn_off_message_logging, | |
mock_response="hello", | |
) | |
await asyncio.sleep(1) | |
standard_logging_payload = test_custom_logger.logged_standard_logging_payload | |
assert standard_logging_payload is not None | |
print( | |
"logged standard logging payload", | |
json.dumps(standard_logging_payload, indent=2), | |
) | |
if turn_off_message_logging is True: | |
assert standard_logging_payload["response"] == {"text": "redacted-by-litellm"} | |
assert ( | |
standard_logging_payload["messages"][0]["content"] == "redacted-by-litellm" | |
) | |
else: | |
assert ( | |
standard_logging_payload["response"]["choices"][0]["message"]["content"] | |
== "hello" | |
) | |
assert standard_logging_payload["messages"][0]["content"] == "hi" | |
async def test_global_redaction_off_with_dynamic_params(turn_off_message_logging): | |
litellm.turn_off_message_logging = False | |
test_custom_logger = TestCustomLogger() | |
litellm.callbacks = [test_custom_logger] | |
response = await litellm.acompletion( | |
model="gpt-3.5-turbo", | |
messages=[{"role": "user", "content": "hi"}], | |
turn_off_message_logging=turn_off_message_logging, | |
mock_response="hello", | |
) | |
await asyncio.sleep(1) | |
standard_logging_payload = test_custom_logger.logged_standard_logging_payload | |
assert standard_logging_payload is not None | |
print( | |
"logged standard logging payload", | |
json.dumps(standard_logging_payload, indent=2), | |
) | |
if turn_off_message_logging is True: | |
assert standard_logging_payload["response"] == {"text": "redacted-by-litellm"} | |
assert ( | |
standard_logging_payload["messages"][0]["content"] == "redacted-by-litellm" | |
) | |
else: | |
assert ( | |
standard_logging_payload["response"]["choices"][0]["message"]["content"] | |
== "hello" | |
) | |
assert standard_logging_payload["messages"][0]["content"] == "hi" | |