LiteLLM allows developers to call all LLM APIs using the openAI format. LiteLLM Proxy is a proxy server to call 100+ LLMs in OpenAI format. Both are supported by this auto-instrumentation.Any calls made to the following functions will be automatically captured by this integration:
Use the register function to connect your application to Phoenix:
Copy
Ask AI
from phoenix.otel import register# configure the Phoenix tracertracer_provider = register( project_name="my-llm-app", # Default is 'default' auto_instrument=True # Auto-instrument your app based on installed OI dependencies)
Add any API keys needed by the models you are using with LiteLLM.
You can now use LiteLLM as normal and calls will be traces in Phoenix.
Copy
Ask AI
import litellmcompletion_response = litellm.completion(model="gpt-3.5-turbo", messages=[{"content": "What's the capital of China?", "role": "user"}])print(completion_response)