Parts of this example uses content from : https://github.com/modelcontextprotocol/quickstart-resources.git - Esp. the weather 'server' code
Code uses two different LLMs just for demonstration. The Proxy Agent uses gpt-4o and the tool uses sonnet. So, export OPENAI_API_KEY as well as ANTHROPIC_API_KEY - OR - modify the code to suit your models
The pyproject.toml assumes you are using 'uv' package manager
- Clone this repo
- uv sync
- cd mcp-client
- uv run client.py (this requires openai and anthropic keys and uses anthropic libs directly)
- uv run client2.py (for pure pydantic and works with any fn calling LLM)
(Alternatively try client2.py - this uses only PydanticAI - no direct dep on Anthropic libs)
Now, try interacting with some questions like:
What is the time in NY when it is 7:30pm in Bangalore?
What is the Weather currently in Chicago?
(and quit when done)