Installation¶
Requirements¶
- Python 3.8 or higher
- pip or uv
Basic Installation¶
Install ullm using pip:
Or using uv (faster):
Optional Dependencies¶
AWS Bedrock Support¶
If you need AWS Bedrock support, install with the aws extra:
This installs boto3 which is required for AWS Bedrock integration.
Development Installation¶
For development, install with dev dependencies:
This includes: - pytest (testing) - ruff (linting and formatting) - mypy (type checking) - pytest-asyncio (async testing) - pytest-cov (coverage)
Documentation¶
To build the documentation locally:
All Optional Dependencies¶
To install everything:
From Source¶
Clone the repository and install in development mode:
Verify Installation¶
Verify your installation:
API Keys¶
ullm reads API keys from environment variables:
# OpenAI
export OPENAI_API_KEY=sk-...
# Anthropic
export ANTHROPIC_API_KEY=sk-ant-...
# Groq
export GROQ_API_KEY=gsk_...
# AWS Bedrock (uses standard AWS credentials)
export AWS_ACCESS_KEY_ID=...
export AWS_SECRET_ACCESS_KEY=...
export AWS_REGION_NAME=us-east-1
Or pass them directly:
import ullm
response = ullm.completion(
model="openai/gpt-4o-mini",
messages=[{"role": "user", "content": "Hi"}],
api_key="sk-..."
)
Next Steps¶
- Quick Start Guide - Get started in 5 minutes
- Basic Usage - Learn the fundamentals
- User Guide - Explore all features