AWS Bedrock Integration
AWS Bedrock integration is available in AxonFlow Enterprise Edition. Full setup documentation including HIPAA-compliant configuration is available in the Enterprise Documentation Portal.
Overview
AWS Bedrock provides serverless access to foundation models from leading AI providers through a unified API. AxonFlow integrates seamlessly with Bedrock, enabling enterprises to leverage multiple LLM providers without managing infrastructure.
Why Use Bedrock with AxonFlow?
- Multi-Model Access: Claude 3.5 Sonnet, Llama 3, Mistral, and more through a single integration
- Serverless: No model hosting infrastructure required
- AWS Native: IAM-based security, CloudWatch monitoring, VPC endpoints
- Cost Effective: Pay-per-token pricing with no minimum commitments
- Compliance Ready: AWS compliance certifications (HIPAA, SOC 2, GDPR)
Architecture
User Request
|
AxonFlow Agent (Governance + Policy Enforcement)
|
AWS Bedrock API (Model Routing)
|
Foundation Models (Claude, Llama, Mistral, etc.)
AxonFlow adds governance and policy enforcement with minimal latency overhead (<10ms P95) while routing requests to Bedrock models.
Supported Models
Anthropic Claude Models (Recommended)
| Model | Model ID | Use Case | Cost (Per 1M Tokens) |
|---|---|---|---|
| Claude 3.5 Sonnet | anthropic.claude-3-5-sonnet-20240620-v1:0 | Complex reasoning, code generation | Input: $3.00, Output: $15.00 |
| Claude 3 Haiku | anthropic.claude-3-haiku-20240307-v1:0 | Fast responses, high throughput | Input: $0.25, Output: $1.25 |
| Claude 3 Opus | anthropic.claude-3-opus-20240229-v1:0 | Highest accuracy, complex tasks | Input: $15.00, Output: $75.00 |
Recommendation: Start with Claude 3.5 Sonnet for balanced performance/cost, switch to Haiku for high-volume use cases.
Meta Llama Models
| Model | Model ID | Use Case | Cost (Per 1M Tokens) |
|---|---|---|---|
| Llama 3 70B | meta.llama3-70b-instruct-v1:0 | Open-source alternative, coding | Input: $0.99, Output: $0.99 |
| Llama 3 8B | meta.llama3-8b-instruct-v1:0 | Budget-friendly, simple tasks | Input: $0.30, Output: $0.30 |
Recommendation: Llama 3 70B for cost-sensitive deployments where Claude's advanced reasoning isn't required.
Mistral AI Models
| Model | Model ID | Use Case | Cost (Per 1M Tokens) |
|---|---|---|---|
| Mistral Large | mistral.mistral-large-2402-v1:0 | European compliance, multilingual | Input: $8.00, Output: $24.00 |
| Mixtral 8x7B | mistral.mixtral-8x7b-instruct-v0:1 | Efficient mixture-of-experts | Input: $0.45, Output: $0.70 |
Recommendation: Mistral for EU data residency requirements or multilingual use cases.
Prerequisites
Before integrating Bedrock with AxonFlow Enterprise, ensure you have:
AWS Account Requirements
- AWS Account with Bedrock enabled
- IAM permissions to enable model access and create policies
- VPC with private subnets (required for HIPAA deployments)
Supported Regions
us-east-1(N. Virginia)us-west-2(Oregon)eu-central-1(Frankfurt)ap-southeast-1(Singapore)
Model Access Approval
Bedrock models require explicit access approval (one-time setup):
- Open AWS Console -> Bedrock -> Model access
- Click "Modify model access"
- Select models you want to enable (Claude 3.5 Sonnet, Claude 3 Haiku, Llama 3 70B)
- Submit access request
- Wait for approval (instant for most models, 1-2 business days for some)
Deployment Options
AWS Marketplace (Recommended)
Deploy AxonFlow Enterprise via AWS Marketplace CloudFormation, which automatically:
- Creates IAM roles with Bedrock permissions
- Configures ECS tasks with appropriate execution roles
- Sets up CloudWatch logging and monitoring
See AWS Marketplace Deployment for details.
Self-Hosted Deployment
For self-hosted deployments, you'll need to:
- Create IAM policy with
bedrock:InvokeModelpermissions - Attach policy to your ECS task execution role or EC2 instance role
- Configure AxonFlow environment variables for Bedrock
Full setup instructions are available in the Enterprise Documentation Portal.
Enterprise Features
The following features require AxonFlow Enterprise:
HIPAA-Compliant Configuration
- VPC endpoint setup for private Bedrock access (traffic never leaves AWS network)
- KMS encryption for CloudWatch logs
- CloudTrail audit logging
- Compliance checklist and verification
Multi-Model Routing
- Intelligent model routing based on query complexity
- Cost-optimized routing policies
- Fallback configuration for high availability
Cost Management
- Budget policies with daily/monthly limits
- Cost tracking per tenant/agent
- Automatic request denial when budget exceeded
Monitoring & Observability
- CloudWatch metrics integration
- Prometheus metrics export
- Cost and latency dashboards
Cost Comparison
Bedrock models are typically 50-90% cheaper than OpenAI equivalents:
| Scenario | OpenAI GPT-4 | Bedrock Claude 3.5 Sonnet | Bedrock Claude 3 Haiku |
|---|---|---|---|
| 1M tokens/month | $40 | $18 (55% savings) | $1.50 (96% savings) |
| 10M tokens/month | $400 | $180 (55% savings) | $15 (96% savings) |
Getting Started
Ready to use AWS Bedrock with AxonFlow?
-
Community Users: AxonFlow Community supports OpenAI and local LLM providers. Get started with Community
-
Enterprise Trial: Contact sales for a 14-day enterprise trial with full Bedrock support. Request Trial
-
AWS Marketplace: Subscribe to AxonFlow Enterprise on AWS Marketplace for immediate deployment.
Related Documentation
FAQ
Which Bedrock model should I use?
- Complex reasoning / code generation: Claude 3.5 Sonnet
- High volume / cost-sensitive: Claude 3 Haiku or Llama 3 70B
- Highest accuracy (worth premium): Claude 3 Opus
- EU compliance / multilingual: Mistral Large
Does AxonFlow add latency to Bedrock calls?
AxonFlow adds <10ms overhead (P95) for governance and policy enforcement.
Is Bedrock HIPAA compliant?
Yes, with AWS BAA signed and VPC endpoints configured. Full HIPAA setup guide available in Enterprise documentation.
Can I use multiple Bedrock models simultaneously?
Yes, AxonFlow Enterprise supports multi-model routing with automatic selection based on query complexity, cost constraints, or custom rules.
Last Updated: December 11, 2025