Overview
Azure OpenAI Service provides access to OpenAI models through Microsoft Azure with enterprise features, compliance, and regional deployment. Model Class:OpenAIChatModel (uses Azure provider)
Authentication
Environment Variables
Using infer_model
Manual Configuration
Examples
Basic Usage
With Custom Deployment
With Managed Identity
With Virtual Network
With Content Filtering
Prompt Caching
Azure OpenAI supports the same prompt caching capabilities as OpenAI:Cache Benefits
- Cost Reduction: 50% savings on cached input tokens
- Latency: 2-4x faster with cache hits
- Automatic: No configuration needed
- Minimum Size: 1024 tokens recommended
Model Parameters
Base Settings
| Parameter | Type | Description | Default |
|---|---|---|---|
max_tokens | int | Maximum tokens to generate | Model-specific |
temperature | float | Sampling temperature (0.0-2.0) | 1.0 |
top_p | float | Nucleus sampling | 1.0 |
seed | int | Random seed | None |
stop_sequences | list[str] | Stop sequences | None |
presence_penalty | float | Token presence penalty | 0.0 |
frequency_penalty | float | Token frequency penalty | 0.0 |
logit_bias | dict[str, int] | Token likelihood modifier | None |
parallel_tool_calls | bool | Allow parallel tools | True |
timeout | float | Request timeout (seconds) | 600 |
Azure-Specific Features
Azure OpenAI includes additional enterprise features:- Content Filtering: Automatic content safety checks
- Managed Identity: Azure AD authentication
- Private Endpoints: VNet integration
- Data Residency: Regional deployment
- Compliance: SOC 2, HIPAA, ISO certifications
Example Configuration
Available Models
Azure OpenAI provides access to OpenAI models through regional deployments:GPT-4o Models
gpt-4o: Latest GPT-4ogpt-4o-mini: Efficient variant
GPT-4 Models
gpt-4: Original GPT-4gpt-4-32k: Extended contextgpt-4-turbo: Faster variant
GPT-3.5 Models
gpt-3.5-turbo: Cost-effectivegpt-3.5-turbo-16k: Extended context
Deployment Configuration
Creating a Deployment
- Create Azure OpenAI resource
- Deploy a model (create deployment)
- Note deployment name (used as model_name)
- Get endpoint and keys
Regional Considerations
Choose region based on:- Data residency requirements
- Latency requirements
- Model availability
- Compliance needs
- East US
- West Europe
- UK South
- Australia East
Best Practices
- Use Managed Identity: More secure than API keys
- Deploy in Same Region: Minimize latency
- Use Private Endpoints: For sensitive data
- Monitor Usage: Track through Azure portal
- Set Up Alerts: For quota and cost management
- Version API Calls: Use specific API versions
- Leverage Content Filtering: Built-in safety features
- Use Deployment Names: Not model IDs
Enterprise Features
Content Filtering
Azure includes multi-level content filtering:- Hate and fairness
- Sexual content
- Violence
- Self-harm
Compliance
Azure OpenAI is compliant with:- SOC 2 Type II
- ISO 27001, 27018, 27701
- HIPAA
- FedRAMP (select regions)
Data Privacy
- Data stays in your Azure tenant
- No data used for model training
- Customer Lockbox support
- Encryption at rest and in transit

