# #### What this tests #### # # This tests if logging to the llmonitor integration actually works # # Adds the parent directory to the system path # import sys # import os # sys.path.insert(0, os.path.abspath("../..")) # from litellm import completion, embedding # import litellm # litellm.success_callback = ["llmonitor"] # litellm.failure_callback = ["llmonitor"] # litellm.set_verbose = True # def test_chat_openai(): # try: # response = completion( # model="gpt-3.5-turbo", # messages=[{"role": "user", "content": "Hi 👋 - i'm openai"}], # user="ishaan_from_litellm" # ) # print(response) # except Exception as e: # print(e) # def test_embedding_openai(): # try: # response = embedding(model="text-embedding-ada-002", input=["test"]) # # Add any assertions here to check the response # print(f"response: {str(response)[:50]}") # except Exception as e: # print(e) # test_chat_openai() # # test_embedding_openai() # def test_llmonitor_logging_function_calling(): # function1 = [ # { # "name": "get_current_weather", # "description": "Get the current weather in a given location", # "parameters": { # "type": "object", # "properties": { # "location": { # "type": "string", # "description": "The city and state, e.g. San Francisco, CA", # }, # "unit": {"type": "string", "enum": ["celsius", "fahrenheit"]}, # }, # "required": ["location"], # }, # } # ] # try: # response = completion(model="gpt-3.5-turbo", # messages=[{ # "role": "user", # "content": "what's the weather in boston" # }], # temperature=0.1, # functions=function1, # ) # print(response) # except Exception as e: # print(e) # # test_llmonitor_logging_function_calling()