Skip to main content

Platform overview

MagOneAI is an enterprise-grade AI agent platform that enables organizations to deploy intelligent agents, automated workflows, integrated tools, and managed knowledge bases securely on Kubernetes infrastructure.
MagOneAI platform overview

Builders

Design agents, workflows, and tools in MagOneAI Studio

Consumers

Run workflows and view results in MagOneAI Hub

Admins

Manage orgs, users, and settings in the Superadmin Portal
All three portals connect to a central API Server + Workflow Orchestrator that handles authentication, RBAC, multi-tenancy, and durable execution. The backend connects to:
  • 12+ Tool Servers — Calendar, Email, Drive, Database, Web Search, and more via MCP
  • Knowledge Base — RAG pipeline with hybrid vector search
  • Secrets Vault — Encrypted credential store (HashiCorp Vault)
  • Self-Hosted LLM — Private GPU inference on your hardware (optional)
  • 3rd-Party LLM — OpenAI, Anthropic, Azure OpenAI APIs

Deployment model

MagOneAI is deployed directly within your cloud environment, ensuring all data remains securely within your infrastructure. MagureLabs manages the deployment remotely via a dedicated control plane, enabling seamless provisioning and operational oversight.
Deployment architecture

How it works

1

MagureLabs control plane

Operated by MagureLabs, the control plane includes:
  • Management Dashboard — Configuration and deployment management
  • Provisioning Engine — Infrastructure-as-Code deployment into your account
  • Health Monitoring — Outbound-only health reports from your cluster
2

Your cloud account

All application workloads run in your account:
  • Kubernetes Cluster — Application services, web portals, tool servers
  • Managed Database — PostgreSQL with HA and automated backups
  • Stateful Services — Cache, Vector DB, Vault, Object Store
  • GPU Pool (optional) — LLM inference, auto-scales 0 to N
MagureLabs uses Infrastructure-as-Code to provision resources and receives only outbound health reports. No inbound access to your cluster is required.

Cloud infrastructure

The following cloud resources are provisioned in your account — networking, compute, managed database, encryption, and storage.
Cloud infrastructure components

Networking

ResourceDescription
Private NetworkVPC + Subnets — all workloads run in private subnets
Load BalancerTLS termination, L7 routing
Firewall RulesRestrictive ingress/egress policies
NAT GatewayOutbound-only internet access
Auto-provisioned TLSLet’s Encrypt certificates, automatically managed
DNS RecordPoints your-domain.com to a static IP (public or intranet)

Compute

ResourceDescription
Kubernetes ClusterManaged, auto-healing, CPU node pool — always on
GPU Node Pool (optional)Auto-scales 0 to N for LLM inference

Managed services

ResourceDescription
Key ManagementCloud-native encryption keys
Secrets ManagerHashiCorp Vault with KMS auto-unseal for encrypted credential storage
PostgreSQLManaged, HA, private networking, automated backups
Object StorageS3-compatible storage for documents and exports

Kubernetes workloads

Inside the cluster, these are the services that make up the platform.
Kubernetes workloads

Web portals

ServiceDescription
StudioAgent builder — design agents, workflows, and tools
HubWorkflow runner — end-user interface for running workflows and viewing results
SuperadminPlatform admin — manage organizations, users, and settings

API and orchestration

ServiceDescription
API ServerAuth, RBAC, multi-tenant API gateway
Workflow EngineTemporal-based durable workflow execution, retries, and scheduling
Workflow WorkersAgent execution — runs workflow activities
Knowledge WorkersDocument ingestion, chunking, and embedding pipeline

Tool servers (12+ integrations)

Google

Google Calendar, Google Gmail

Microsoft

Outlook Calendar, Outlook Email, OneDrive

Data

Database (Text-to-SQL), Web Search

Utilities

Filesystem, File Tools, Checkbox Detector

Stateful services

ServiceDescription
Vector DBQdrant — RAG semantic search
Object StoreS3-compatible document and file storage
CacheRedis — sessions, pub/sub
Secrets VaultHashiCorp Vault with KMS auto-unseal

GPU pool (optional)

ServiceDescription
LLM Inference ServerSelf-hosted model serving, scales from zero

LLM hosting

MagOneAI offers flexible hosting options tailored to your privacy and cost requirements, with the ability to combine multiple approaches for a hybrid deployment model.
LLM hosting options
Run open-source models on your own GPU hardware. Data never leaves your infrastructure.
GPU nodesRequired
ModelsOpen-source (Llama, Mistral, etc.) run on your hardware
Data privacyData never leaves your infrastructure
ScalingGPU auto-scales from zero
Air-gappedSupports fully air-gapped deployment with zero internet access
Self-hosted LLM deployments enable fully air-gapped operations with zero internet dependency, supporting both public and internal/intranet entry points to align with your network security requirements.

Network security

All infrastructure is designed with defense-in-depth security principles.
Network security architecture

Security guarantees

Private network architecture

All infrastructure nodes operate within private subnets without public IP addresses, ensuring network isolation

Unidirectional connectivity

System maintains outbound-only communication protocols, preventing any inbound access or data push to your cluster environment

Comprehensive encryption

Data protection includes encryption at rest through cloud-native key management services and automatic TLS encryption for all data in transit

Flexible access control

Entry point configuration supports both public IP and internal/intranet IP options, allowing you to define access based on your specific security requirements

Air-gapped capability

Self-hosted LLM configurations enable completely isolated deployments with zero internet connectivity for maximum security compliance

Zero-trust architecture

The platform implements outbound-only connectivity patterns, ensuring no external systems can initiate connections to your infrastructure

Network architecture

Traffic flow through the infrastructure:
1

Users connect via HTTPS

Users access the platform through the load balancer. HTTPS only (port 443) with TLS termination.
2

Load balancer routes to cluster

The load balancer sits inside your cloud firewall boundary and routes to the Kubernetes cluster.
3

Private subnet isolation

Application pods run in a private subnet with no public IPs. Network policies enforce pod-to-pod restrictions.
4

Outbound via NAT Gateway

Outbound HTTPS traffic (e.g., 3rd-party LLM API calls) routes through the NAT Gateway. No inbound connections from the internet reach the pods directly.
5

MagureLabs — pull-based updates

MagureLabs delivers updates via pull-based mechanisms. No inbound access to your cluster is needed.

Internal service security

ComponentSecurity measures
Application PodsNetwork policies enforced, pod-to-pod restrictions
Secrets VaultCloud KMS auto-unseal, encrypted at rest
DatabasePrivate IP only, encrypted at rest, automated backups
Container RegistryPrivate registry with certificate authority

Deployment prerequisites

To deploy MagOneAI Enterprise Edition, we need the following from your team:
RequirementDetails
Cloud account accessCredentials to provision: K8s cluster, database, networking, encryption keys
Domain + DNSA subdomain and ability to create an A record
GPU quota (only if self-hosting LLM)We provide exact quota details based on your model selection
MagureLabs handles the entire provisioning and deployment process. Your team provides cloud account access and a domain — we take care of the rest.

Next steps

Security overview

Deep dive into MagOneAI’s security architecture

Secrets management

How Vault manages API keys, OAuth tokens, and credentials

Private models

Configure self-hosted LLM inference

Organizations and projects

Multi-tenant organizational structure