Deploy the full AI stack on your infrastructure—air-gapped, zero external calls, infrastructure activity journal. Sovereignty enforced at the architecture layer.
Capabilities
Foundational blocks to run any generative AI workload inside your sovereign perimeter.
Ingest millions of documents into a sovereign vector store. Connect SharePoint, OneDrive, file shares, or custom sources.
Agents that analyze reports, process forms, and summarize contracts at scale.
Host an on-prem copilot fine-tuned on your repos—no IP leakage.
Query SQL and warehouses in natural language to generate insights and charts.
Architecture
A zero-trust request path from gateway to the infrastructure activity journal.
Components
Five components, one sovereign platform.
Unified API gateway that routes prompts by cost, latency, and sovereignty. Handles auth, rate limits, and orchestration.

Isolation layer enforcing mTLS, activity journaling, and data residency at the infrastructure level.

Optimized inference for MENA workloads on A100/H100 with vLLM and high concurrency.

Air-gapped and private-cloud deployments with offline bundles and immutable rollout tracking.

Web console for models, costs, traces, and fine-tuning with full RTL and Arabic UI.

Operational Assurance
Operational controls that keep sovereignty measurable and predictable.
Outbound traffic is blocked by default; exceptions require operator approval and activity journaling.
Append-only activity journal retained locally per your policy.
Approved model registry with version pinning and rollback across environments.
Per-team budgets, token caps, and automated routing to stay within spend.
Deployment
Air-gapped, sovereign private cloud, or hybrid managed—choose what fits.
Complete isolation from the public internet. Updates are delivered via secure physical media or local registry. Ideal for Defense & National Security.
Deployed within your VPC (AWS Outposts, Azure Stack, or Local Data Center). Data never leaves your controlled perimeter.
Atlas Studio assisted by MX4; data stays in your infrastructure. Balanced operational ease and data security.
Infrastructure
Reference hardware profiles for dev, POC, and production at scale.
| Profile | CPU | RAM | GPU | Storage |
|---|---|---|---|---|
| Development (1 Node) | 4 vCPU | 16 GB | None (CPU only) | 100 GB NVMe |
| Proof of Concept (2 Nodes) | 16 vCPU | 64 GB | 2x NVIDIA A100 (40GB) | 500 GB NVMe |
| Production HA (3+ Nodes) | 32+ vCPU | 256+ GB | 4-8x NVIDIA A100 (80GB) | 2 TB+ NVMe RAID |
Performance
Atlas leads Arabic tasks while running entirely on your infrastructure.
| Metric | MX4 Atlas (Falcon-7B) | GPT-4 (OpenAI) | Improvement | Context |
|---|---|---|---|---|
| Arabic QA Accuracy | 91.2% | 76.8% | +14.4% | Falcon-7B vs GPT-4 on SQuAD-style Arabic dataset |
| Cultural Context | 94.7% | 68.2% | +26.5% | Understanding of local idioms, laws, and customs |
| Dialect Handling | 89.3% | 72.1% | +17.2% | Performance on MSA, Gulf, and Levantine dialects |
| Inference Speed (Tokens/s) | 45 t/s | 18 t/s | 2.5x Faster | Falcon-7B vs GPT-4 API response time |
Sovereignty
Infrastructure-level isolation plus data residency controls you can verify.
Scenarios
Representative examples of how sovereign teams deploy Atlas.
Government
“Atlas Runtime keeps Arabic document workflows on‑prem with zero external calls and infrastructure activity journaling.”
Digital transformation team
Outcome: Faster approvals
Financial Services
“Atlas Core routes requests across local and premium models to balance cost and latency while staying on‑prem.”
AI innovation team
Outcome: Faster onboarding
Healthcare
“Clinical support workflows run inside the hospital perimeter with strict access controls and local models.”
Clinical informatics team
Outcome: Operational efficiency
Critical Infrastructure
“Air‑gapped deployments enable offline inference with sealed update bundles.”
Security architecture team
Outcome: Offline operations
FAQ
Answers for procurement, deployment, and security teams.
Yes. Atlas Deploy supports full offline deployment with bundled models, dependencies, and local container registries. Updates are delivered via secure physical media or dedicated secure channels. No outbound internet connection required.
Atlas supports all major open-source models (Llama, Falcon, Jais, Qwen, Mistral) and can route to OpenAI/Anthropic APIs when appropriate. You can also fine-tune models on your data and host them through Atlas Serve.
Dev environments come online in ~10 minutes. The Free POC Pilot typically installs in ~30 minutes. Production deployments complete in under ~2 hours (including infrastructure checks); broader rollout timing depends on internal readiness.
Atlas is optimized for Modern Standard Arabic and major dialects (Gulf, Levantine, Egyptian). We route Arabic queries to native models like JAIS-13B and Falcon-7B which outperform GPT-4 on regional context by 25%+.
Yes. Atlas integrates with Kubernetes, supports standard OIDC/SAML authentication, works with existing observability stacks (Prometheus, Grafana), and provides OpenAI-compatible APIs for seamless migration.
Atlas focuses on infrastructure. Governance and policy enforcement live in your own stack; Atlas delivers residency controls, isolation, and activity journaling by design.
Yes. Atlas is designed for incremental deployment: start with a POC (30 days, limited scale), validate use cases, then deploy production workloads on your infrastructure. The API remains identical across environments.
Enterprise customers get dedicated support channels, 4-hour response SLA for critical issues, quarterly business reviews, and direct access to our engineering team for architecture guidance.
Get Started
Launch a free POC on your infrastructure. 30 days, zero external dependencies.
Start with a 30-day POC • Typical deployment: 2–4 weeks