MX4 Atlas orchestrates AI infrastructure on your servers – Arabic-native, any model, any cloud, zero vendor lock-in.
Capabilities
Foundational blocks to run any generative AI workload inside your sovereign perimeter.
Ingest millions of documents into a sovereign vector store. Connect SharePoint, OneDrive, file shares, or custom sources.
Agents that analyze reports, process forms, and summarize contracts at scale.
Host an on-prem copilot fine-tuned on your repos—no IP leakage.
Query SQL and warehouses in natural language to generate insights and charts.
Architecture
A zero-trust request path from gateway to the infrastructure activity journal.
Components
Five components, one sovereign platform.
Unified API gateway that routes prompts by cost, latency, and sovereignty. Handles auth, rate limits, and orchestration.

Isolation layer enforcing mTLS, activity journaling, and data residency at the infrastructure level.

Optimized inference for regional workloads on modern GPUs with high concurrency.

Air-gapped and private-cloud deployments with offline bundles and immutable rollout tracking.

Web console for models, costs, traces, and fine-tuning with Arabic-ready UI.

Operational Assurance
Operational controls that keep sovereignty measurable and predictable.
Outbound traffic is blocked by default; exceptions require operator approval and activity journaling.
Append-only activity journal retained locally per your policy.
Approved model registry with version pinning and rollback across environments.
Per-team budgets, token caps, and automated routing to stay within spend.
Deployment
Air-gapped, sovereign private cloud, or hybrid managed—choose what fits.
Complete isolation from the public internet. Updates are delivered via secure physical media or local registry. Ideal for Defense & National Security.
Deployed within your VPC (AWS Outposts, Azure Stack, or Local Data Center). Data never leaves your controlled perimeter.
Atlas Studio assisted by MX4; data stays in your infrastructure. Balanced operational ease and data security.
Infrastructure
Example reference profiles for dev, POC, and production at scale.
| Profile | CPU | RAM | GPU | Storage |
|---|---|---|---|---|
| Development (1 Node) | 4 vCPU | 16 GB | None (CPU only) | 100 GB NVMe |
| Proof of Concept (2 Nodes) | 16 vCPU | 64 GB | 2x data center GPUs (A100-class) | 500 GB NVMe |
| Production HA (3+ Nodes) | 32+ vCPU | 256+ GB | 4-8x data center GPUs (A100-class) | 2 TB+ NVMe RAID |
Performance
Indicative internal evaluations while running entirely on your infrastructure.
| Metric | MX4 Atlas | Reference Cloud | Relative | Context |
|---|---|---|---|---|
| Arabic QA Accuracy | Strong | Baseline | Higher | Internal evaluation on Arabic QA tasks |
| Cultural Context | High | Baseline | Higher | Regional context and idiom handling |
| Dialect Handling | High | Baseline | Higher | MSA, Gulf, and Levantine dialects |
| Inference Speed (Tokens/s) | Low latency | Baseline | Faster | Local GPU inference vs external API |
Sovereignty
Infrastructure-level isolation plus data residency controls you can verify.
Scenarios
Representative examples of how sovereign teams deploy Atlas.
Government
“Atlas Runtime keeps Arabic document workflows on‑prem with zero external calls and infrastructure activity journaling.”
Digital transformation team
Outcome: Faster approvals
Financial Services
“Atlas Core routes requests across local and premium models to balance cost and latency while staying on‑prem.”
AI innovation team
Outcome: Faster onboarding
Healthcare
“Clinical support workflows run inside the hospital perimeter with strict access controls and local models.”
Clinical informatics team
Outcome: Operational efficiency
Critical Infrastructure
“Air‑gapped deployments enable offline inference with sealed update bundles.”
Security architecture team
Outcome: Offline operations
FAQ
Answers for procurement, deployment, and security teams.
Yes. Atlas Deploy supports full offline deployment with bundled models, dependencies, and local container registries. Updates are delivered via secure physical media or dedicated secure channels. No outbound internet connection required.
Atlas supports major open-source models (Llama, Falcon, Jais, Qwen, Mistral) and can optionally route to external APIs if you choose. You can also fine-tune models on your data and host them through Atlas Serve.
Dev environments come online in minutes. The Free POC Pilot typically installs the same day. Production timelines depend on infrastructure readiness and security reviews.
Atlas is optimized for Modern Standard Arabic and major dialects (Gulf, Levantine, Egyptian). We route Arabic queries to native models and validate performance with internal evaluations.
Yes. Atlas integrates with Kubernetes, supports standard OIDC/SAML authentication, works with existing observability stacks (Prometheus, Grafana), and provides OpenAI-compatible APIs for seamless migration.
Atlas focuses on infrastructure. Governance and policy enforcement live in your own stack; Atlas delivers residency controls, isolation, and activity journaling by design.
Yes. Atlas is designed for incremental deployment: start with a POC (30 days, limited scale), validate use cases, then deploy production workloads on your infrastructure. The API remains identical across environments.
Enterprise customers get dedicated support channels, response SLAs for critical issues, and direct access to our engineering team for architecture guidance.
Get Started
Launch a free POC on your infrastructure. 30 days, zero external dependencies.
Start with a 30-day POC • Typical deployment: 2–4 weeks