Skip to content

Self-hosted AI coding data boundary

MergeLoom is designed around a customer-hosted worker boundary for self-hosted AI coding.

The worker handles the sensitive execution path. The control plane coordinates the workflow.

MergeLoom architecture diagram showing the customer-hosted worker boundary and controller coordination path.
The customer controller coordinates queue state and review output. The worker performs checkout, context assembly, model calls, tests, and branch preparation inside your infrastructure.

The worker performs:

  • repository checkout
  • context assembly
  • AI execution path
  • command execution
  • validation
  • repair attempts
  • branch preparation
  • local live streams and deeper run traces

This means the most sensitive operational activity happens inside your environment.

MergeLoom worker UI showing local worker runs inside the customer-hosted boundary.
Local worker UI: run streams and trace detail live with the worker, close to repository checkout, model calls, and validation commands.

The control plane coordinates:

  • tenant/workspace setup
  • integration connection state
  • repository catalog metadata
  • workflow rules
  • job envelopes
  • worker enrollment
  • review request creation/update coordination
  • visible run state

The control plane should not be treated as the place where repository execution happens.

Model calls are made from the worker path.

That matters because your provider choice and network policy can be aligned with your own infrastructure rules.

Provider readiness and credentials are managed in the local worker UI, usually http://127.0.0.1:8010/ after installation. If the worker is on a VPS, open the worker UI through an SSH tunnel to port 8010.

Examples:

  • use Codex CLI from the worker container
  • use Claude Code CLI from the worker container
  • point the worker at a private OpenAI-compatible endpoint
  • use approved cloud provider endpoints from inside your network

Before rollout, review:

  • where the worker runs
  • which repositories the worker can access
  • which commands the worker can run
  • which provider endpoint the worker can call
  • which secrets are mounted into the worker
  • who can change provider and runtime settings
  • how long worker-local traces are retained
  • how PR/MR output is approved

For Kubernetes installs, prefer secret.existingSecretName and a cluster-managed Secret for enrollment tokens and provider API keys. Mount only the keys the worker needs, such as JCA_WORKER_ENROLLMENT_TOKEN, JCA_OPENAI_API_KEY, or provider-specific cloud credentials.

For cloud AI providers, prefer workload identity or IAM roles instead of static keys when possible. That means GKE Workload Identity Federation for Vertex AI, EKS IRSA or another AWS role source for Bedrock, and AKS Workload Identity or managed identity for Azure Foundry.

For a first security review:

  1. Use a non-production repository.
  2. Use one worker.
  3. Use one provider.
  4. Use explicit allowed commands.
  5. Use clear validation commands.
  6. Keep human review required before merge.
  7. Confirm logs and traces contain what your reviewers need.