Skip to main content

AWS Bedrock Integration

Enterprise Feature

AWS Bedrock integration is available in AxonFlow Enterprise Edition. Full setup documentation including HIPAA-compliant configuration is available in the Enterprise Documentation Portal.

Contact Sales | Request Demo

Overview

AWS Bedrock provides serverless access to foundation models from leading AI providers through a unified API. AxonFlow integrates seamlessly with Bedrock, enabling enterprises to leverage multiple LLM providers without managing infrastructure.

Why Use Bedrock with AxonFlow?

  • Multi-Model Access: Claude 3.5 Sonnet, Llama 3, Mistral, and more through a single integration
  • Serverless: No model hosting infrastructure required
  • AWS Native: IAM-based security, CloudWatch monitoring, VPC endpoints
  • Cost Effective: Pay-per-token pricing with no minimum commitments
  • Compliance Ready: AWS compliance certifications (HIPAA, SOC 2, GDPR)

Architecture

User Request
|
AxonFlow Agent (Governance + Policy Enforcement)
|
AWS Bedrock API (Model Routing)
|
Foundation Models (Claude, Llama, Mistral, etc.)

AxonFlow adds governance and policy enforcement with minimal latency overhead (<10ms P95) while routing requests to Bedrock models.


Supported Models

ModelModel IDUse CaseCost (Per 1M Tokens)
Claude 3.5 Sonnetanthropic.claude-3-5-sonnet-20240620-v1:0Complex reasoning, code generationInput: $3.00, Output: $15.00
Claude 3 Haikuanthropic.claude-3-haiku-20240307-v1:0Fast responses, high throughputInput: $0.25, Output: $1.25
Claude 3 Opusanthropic.claude-3-opus-20240229-v1:0Highest accuracy, complex tasksInput: $15.00, Output: $75.00

Recommendation: Start with Claude 3.5 Sonnet for balanced performance/cost, switch to Haiku for high-volume use cases.

Meta Llama Models

ModelModel IDUse CaseCost (Per 1M Tokens)
Llama 3 70Bmeta.llama3-70b-instruct-v1:0Open-source alternative, codingInput: $0.99, Output: $0.99
Llama 3 8Bmeta.llama3-8b-instruct-v1:0Budget-friendly, simple tasksInput: $0.30, Output: $0.30

Recommendation: Llama 3 70B for cost-sensitive deployments where Claude's advanced reasoning isn't required.

Mistral AI Models

ModelModel IDUse CaseCost (Per 1M Tokens)
Mistral Largemistral.mistral-large-2402-v1:0European compliance, multilingualInput: $8.00, Output: $24.00
Mixtral 8x7Bmistral.mixtral-8x7b-instruct-v0:1Efficient mixture-of-expertsInput: $0.45, Output: $0.70

Recommendation: Mistral for EU data residency requirements or multilingual use cases.


Prerequisites

Before integrating Bedrock with AxonFlow Enterprise, ensure you have:

AWS Account Requirements

  • AWS Account with Bedrock enabled
  • IAM permissions to enable model access and create policies
  • VPC with private subnets (required for HIPAA deployments)

Supported Regions

  • us-east-1 (N. Virginia)
  • us-west-2 (Oregon)
  • eu-central-1 (Frankfurt)
  • ap-southeast-1 (Singapore)

Model Access Approval

Bedrock models require explicit access approval (one-time setup):

  1. Open AWS Console -> Bedrock -> Model access
  2. Click "Modify model access"
  3. Select models you want to enable (Claude 3.5 Sonnet, Claude 3 Haiku, Llama 3 70B)
  4. Submit access request
  5. Wait for approval (instant for most models, 1-2 business days for some)

Deployment Options

Deploy AxonFlow Enterprise via AWS Marketplace CloudFormation, which automatically:

  • Creates IAM roles with Bedrock permissions
  • Configures ECS tasks with appropriate execution roles
  • Sets up CloudWatch logging and monitoring

See AWS Marketplace Deployment for details.

Self-Hosted Deployment

For self-hosted deployments, you'll need to:

  1. Create IAM policy with bedrock:InvokeModel permissions
  2. Attach policy to your ECS task execution role or EC2 instance role
  3. Configure AxonFlow environment variables for Bedrock

Full setup instructions are available in the Enterprise Documentation Portal.


Enterprise Features

The following features require AxonFlow Enterprise:

HIPAA-Compliant Configuration

  • VPC endpoint setup for private Bedrock access (traffic never leaves AWS network)
  • KMS encryption for CloudWatch logs
  • CloudTrail audit logging
  • Compliance checklist and verification

Multi-Model Routing

  • Intelligent model routing based on query complexity
  • Cost-optimized routing policies
  • Fallback configuration for high availability

Cost Management

  • Budget policies with daily/monthly limits
  • Cost tracking per tenant/agent
  • Automatic request denial when budget exceeded

Monitoring & Observability

  • CloudWatch metrics integration
  • Prometheus metrics export
  • Cost and latency dashboards

Cost Comparison

Bedrock models are typically 50-90% cheaper than OpenAI equivalents:

ScenarioOpenAI GPT-4Bedrock Claude 3.5 SonnetBedrock Claude 3 Haiku
1M tokens/month$40$18 (55% savings)$1.50 (96% savings)
10M tokens/month$400$180 (55% savings)$15 (96% savings)

Getting Started

Ready to use AWS Bedrock with AxonFlow?

  1. Community Users: AxonFlow Community supports OpenAI and local LLM providers. Get started with Community

  2. Enterprise Trial: Contact sales for a 14-day enterprise trial with full Bedrock support. Request Trial

  3. AWS Marketplace: Subscribe to AxonFlow Enterprise on AWS Marketplace for immediate deployment.



FAQ

Which Bedrock model should I use?

  • Complex reasoning / code generation: Claude 3.5 Sonnet
  • High volume / cost-sensitive: Claude 3 Haiku or Llama 3 70B
  • Highest accuracy (worth premium): Claude 3 Opus
  • EU compliance / multilingual: Mistral Large

Does AxonFlow add latency to Bedrock calls?

AxonFlow adds <10ms overhead (P95) for governance and policy enforcement.

Is Bedrock HIPAA compliant?

Yes, with AWS BAA signed and VPC endpoints configured. Full HIPAA setup guide available in Enterprise documentation.

Can I use multiple Bedrock models simultaneously?

Yes, AxonFlow Enterprise supports multi-model routing with automatic selection based on query complexity, cost constraints, or custom rules.


Last Updated: December 11, 2025