Deploy Collate AI Proxy (CAIP)
Complete Deploy Collate before this step. CAIP must run in the same namespace as Collate.
LLM Provider Support
| Provider | Status | Chat Models | Embedding Models |
|---|---|---|---|
| AWS Bedrock | ✅ Default & Recommended | Sonnet 4.5, Haiku 4.5 | Amazon Titan |
| Azure OpenAI | Available (1.12+) | GPT-4o | text-embedding-3-small |
| OpenAI | Available (1.12+) | GPT-4o | text-embedding-3-small |
IAM Permissions for Bedrock (ROSA Only)
This section is ROSA/AWS-specific. For Azure OpenAI, skip to Configure Helm Values.
openmetadata) and its existing IRSA role (openmetadata-rosa-role). Add a Bedrock inline policy to that role:
Configure Helm Values
Create avalues-caip-openshift.yaml file for your provider.
AWS Bedrock
Azure OpenAI
For Azure OpenAI setup (creating a resource group, deploying models, and obtaining credentials), see the Azure OpenAI Configuration section below.
Deploy
Add the Helm repo and deploy CAIP into theopenmetadata namespace:
Verify the Pod
Configure Collate to Use CAIP
After CAIP is running, update your Collate deployment to point to it.Collate Helm Chart
Add the following to yourvalues-openshift.yaml and run helm upgrade:
Environment Variables (Custom Chart)
If you deployed Collate with a custom Helm chart, add these environment variables: AWS Bedrock:Validation
Once deployed and configured, verify CAIP is healthy from the Collate UI: Settings → Preferences → Health Check All AI services should show a healthy status.Hardware Requirements
CAIP is stateless — no persistent storage is required.| Resource | Minimum |
|---|---|
| CPU | 300m |
| Memory | 2Gi |
| Storage | None |
Azure OpenAI Configuration
If using Azure OpenAI, follow these steps to create the required resources:- Sign in to the Azure portal and create a dedicated resource group.
- Navigate to Azure OpenAI (Microsoft Foundry) and create a new service. Fill in the resource group, name, and region; leave other settings as default.
- Once created, click Go to Foundry Portal (Microsoft Foundry).
- Under Shared Resources → Deployments, click Deploy model → Deploy base model.
- Deploy a chat model (e.g.
gpt-4o) for CAIP. - Deploy a separate embedding model (
text-embedding-3-small) for Collate Server semantic search.
- Deploy a chat model (e.g.
- From each deployment’s detail page, collect: API key, deployment name, API version, base URL, and resource name.
Supported Models
All OpenAI GPT models are supported. Collate recommends gpt-4o for CAIP.Troubleshooting
Pod Stuck in Pending
| Symptom | Cause | Fix |
|---|---|---|
Insufficient cpu / memory | Cluster at capacity | Reduce resources.requests in values-caip-openshift.yaml |
ImagePullBackOff | ECR pull secret missing or expired | Recreate collate-ecr-secret — see Deploy Collate |
unable to validate against any SCC | Security context incompatible with restricted-v2 | Ensure podSecurityContext: {} and securityContext block match the values above |
Bedrock Access Denied (ROSA)
- Confirm the
bedrock-accessinline policy was attached:aws iam get-role-policy --role-name openmetadata-rosa-role --policy-name bedrock-access - Verify the trust policy
subcondition matchessystem:serviceaccount:openmetadata:openmetadataexactly. - Confirm the Anthropic models are enabled in your AWS region in the Bedrock console.
CAIP Cannot Reach Collate
CAIP connects to Collate via the Kubernetes service DNS name. Test connectivity from inside the CAIP pod:{"status":"OK"}
If this fails, check that hostAndPort in your values matches the actual Collate service name and port in the openmetadata namespace: