|
|
|
|
|
|
|
import sys, os |
|
import traceback |
|
import pytest |
|
|
|
sys.path.insert( |
|
0, os.path.abspath("../..") |
|
) |
|
import litellm, asyncio |
|
|
|
|
|
@pytest.mark.asyncio |
|
async def test_azure_health_check(): |
|
response = await litellm.ahealth_check( |
|
model_params={ |
|
"model": "azure/chatgpt-v-2", |
|
"messages": [{"role": "user", "content": "Hey, how's it going?"}], |
|
"api_key": os.getenv("AZURE_API_KEY"), |
|
"api_base": os.getenv("AZURE_API_BASE"), |
|
"api_version": os.getenv("AZURE_API_VERSION"), |
|
} |
|
) |
|
print(f"response: {response}") |
|
|
|
assert "x-ratelimit-remaining-tokens" in response |
|
return response |
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio |
|
async def test_azure_embedding_health_check(): |
|
response = await litellm.ahealth_check( |
|
model_params={ |
|
"model": "azure/azure-embedding-model", |
|
"api_key": os.getenv("AZURE_API_KEY"), |
|
"api_base": os.getenv("AZURE_API_BASE"), |
|
"api_version": os.getenv("AZURE_API_VERSION"), |
|
}, |
|
input=["test for litellm"], |
|
mode="embedding", |
|
) |
|
print(f"response: {response}") |
|
|
|
assert "x-ratelimit-remaining-tokens" in response |
|
return response |
|
|
|
|
|
@pytest.mark.asyncio |
|
async def test_openai_img_gen_health_check(): |
|
response = await litellm.ahealth_check( |
|
model_params={ |
|
"model": "dall-e-3", |
|
"api_key": os.getenv("OPENAI_API_KEY"), |
|
}, |
|
mode="image_generation", |
|
prompt="cute baby sea otter", |
|
) |
|
print(f"response: {response}") |
|
|
|
assert isinstance(response, dict) and "error" not in response |
|
return response |
|
|
|
|
|
|
|
|
|
|
|
async def test_azure_img_gen_health_check(): |
|
response = await litellm.ahealth_check( |
|
model_params={ |
|
"model": "azure/", |
|
"api_base": os.getenv("AZURE_API_BASE"), |
|
"api_key": os.getenv("AZURE_API_KEY"), |
|
"api_version": "2023-06-01-preview", |
|
}, |
|
mode="image_generation", |
|
prompt="cute baby sea otter", |
|
) |
|
|
|
assert isinstance(response, dict) and "error" not in response |
|
return response |
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.asyncio |
|
async def test_sagemaker_embedding_health_check(): |
|
response = await litellm.ahealth_check( |
|
model_params={ |
|
"model": "sagemaker/berri-benchmarking-gpt-j-6b-fp16", |
|
"messages": [{"role": "user", "content": "Hey, how's it going?"}], |
|
}, |
|
mode="embedding", |
|
input=["test from litellm"], |
|
) |
|
print(f"response: {response}") |
|
|
|
assert isinstance(response, dict) |
|
return response |
|
|
|
|
|
|
|
|