Production-ready · Built in Go & Rust · Kubernetes-native

AI orchestration built to ship fast and scale hard.

Code AI workflows fast.
Run production AI without compromises.

All nodes run in Go, with Go and Rust streaming workers on Kubernetes. Client and user isolation is built in. Per-user memory stays persistent. Workflows can mutate live when performance shifts. Each node autoscales in isolation. From first workflow to enterprise load, one runtime.

400+ integrations
E2EE by default
Per-node autoscaling
Multi-tenant isolation
textcorrection.mermaid 4 lines · encrypted
flowchart TD
    A[Progress:40:processing] --> B
    B{{OpenAI:chat}} *model=claude-4.6* --> C
    C[Progress:80:post_processing] --> D
    D((Notify:correction)) *e2ee=auto*
Executing · SSE streaming active

Built different. By design.

Every feature is a structural advantage. Not a checkbox.

01

Mermaid DSL

Define entire AI pipelines in human-readable syntax. LLM-native, trivially mutable. An AI can generate, modify, and reason about workflows as easily as reading a sentence.

B{{Agent:ProcessText}} *model=@best*
02

Per-node Kubernetes scaling

Your 2ms transform and 5-second LLM inference don't share the same scaling policy. Each node gets its own HPA, its own resource profile. 1 pod for transforms. 8 for inference bursts.

Node A: 1 replica -> Node B: 8 replicas
03

Zero-knowledge encryption

ChaCha20-Poly1305 AEAD with one flag. The encryption key is never stored by ionun and is provided externally at runtime. Even if storage is compromised, user data stays unreadable.

D((Notify:result)) *e2ee=auto*
04

Per-user persistent memory

Every user gets dedicated fast memory: isolated, hot-cached, encrypted at rest. Workflows remember conversation history, preferences, accumulated context across sessions.

Redis-backed · zero-knowledge · auto-tiered
05

Adaptive workflows

When a node underperforms (2x latency, rising error rate), ionun can branch, swap, alert, or predict. Autonomous orchestration: workflows that heal and optimize themselves.

model=@best -> auto-failover on degradation
06

Native security DSL

Credentials injected from secret stores at runtime. Multi-tenant isolation at the database level. Client auth via API key + short-lived JWT. Security in the syntax, not as an afterthought.

e2ee=auto · secret_store=vault · jwt=true
How it works

Write workflows in Mermaid.

No drag-and-drop builder. No JSON walls. Just syntax that humans and LLMs understand natively.

analysis-pipeline.mermaid 6 lines · 2 agents
flowchart TD
    A{SSE:ANALYSIS_STARTED} *data=$analysisId* --> B
    B(Transform:AssembleText) --> C
    C{{Agent:ProcessText}} *model=claude-4.6 timeout=3m input=$text* --> D
    D[(Fetch:GetUserData)] --> E
    E{{Agent:MatchResults}} *model=@best input=$C.result,$D.data* --> F
    F{SSE:ANALYSIS_COMPLETED}
This pipeline
  1. Emits a real-time SSE event
  2. Assembles input data
  3. Runs an AI agent on the text
  4. Fetches user data from your backend
  5. Runs a second AI agent cross-referencing both outputs
  6. Signals completion via SSE
6 lines. Two AI agents. Real-time streaming. Cross-node data flow.

Node metadata

B{{Agent:ProcessText}} *model=claude-4.6 timeout=3m input=$assembled_text*
Key What it does
model=claude-4.6 Explicit model selection
model=@best ionun picks the optimal model in real-time
model=@fastest Optimize for latency
timeout=3m Execution time limit
input=$text Pull from workflow context
input=$C.result Reference another node's output
e2ee=auto ChaCha20-Poly1305 end-to-end encryption
streaming=true Enable streaming output
Per-user memory

AI products need memory,
not just execution.

ionun gives every user their own persistent memory: fast, isolated, hot-cached, and encrypted at rest. The encryption key is never stored by ionun.

When your user comes back tomorrow, ionun already knows them. This turns a stateless pipeline runner into a runtime for intelligent products.

  • Isolated per client Each product has its own namespace
  • Per-user fast state Hot Redis-backed memory per user
  • Encrypted at rest Key never stored by ionun
  • Zero-knowledge Key provided externally at runtime
  • Auto-tiered Hot to cold storage automatically
  • Context-aware History, preferences, accumulated data
ionun Memory Layer
Client A
User 1
Memory fast · encrypted
Client B
User 1
Memory fast · enc.
User 2
Memory fast · enc.
Redis (hot)
auto-tier
Cold storage
Zero-knowledge · Key never stored by ionun
Per-Node Independent Scaling
Node A
light
1 replica
Standard HPA
->
Node B
CPU-heavy
8 replicas
Burst HPA
->
Node C
GPU inference
4 replicas
GPU-aware HPA
Per-node scaling

Scaling that matches reality.

Your AI pipeline has a 2ms data transform and a 5-second LLM inference in the same workflow. They don't need the same resources. They shouldn't share the same scaling policy.

ionun scales each node independently on Kubernetes with separate replicas, separate HPAs, separate resource profiles. No other orchestration engine does this.

Independent HPAs Each node type has its own autoscaling policy
Resource isolation CPU, memory, GPU profiles per node
Burst scaling Inference nodes burst on demand, independently

400+ nodes. 60+ categories.

Everything you need, ready to wire into any workflow.

AI
OpenAIAnthropicGroqMistralOllamaHuggingFace
LangChain
ChainsAgentsMemoryRetrieversTools
VectorStore
PineconeQdrantWeaviateChromaMilvus
Database
PostgreSQLMySQLMongoDBSupabaseRedis
Communication
SlackDiscordTelegramWhatsAppEmail
Storage
S3GCSAzure BlobDropboxDrive
DevOps
GitHubGitLabJenkinsJiraKubernetes
Security
OAuthJWTHMACEncryptionKey Mgmt
Finance
StripePayPalWiseAccounting
CRM
HubSpotSalesforcePipedriveActiveCampaign
Marketing
MailchimpSendGridActiveCampaign
Monitoring
DatadogPagerDutyGrafanaPrometheus
+ 48 more categories including Analytics, IoT, Healthcare, Legal, HR, and more
Architecture

Enterprise AI Infrastructure

7 microservices. Go backend. Svelte dashboard. PostgreSQL. Redis. Kubernetes.

Clients SDK / API
Gateway REST · SSE · WebSocket · Auth · E2EE
ionun Core History · Matcher · Mermaid Parser
PostgreSQL
Workers
Go · AI Go · Data Rust · Optim
enterprise-agent-pipeline.mermaid
flowchart TD
    A[(Fetch:UserContext)] *input=$userId* --> B
    B{{Agent:Understand}} *model=@fastest input=$A.context,$message* --> C
    C{Switch:intent} *mode=rules*
    C --> |query| D{{Agent:Retrieve}} *model=claude-4.6*
    C --> |action| E{{Agent:Execute}} *model=@best*
    C --> |memory| F(Transform:UpdateMemory)
    D --> G((Return:response))
    E --> G
    F --> G
What this gives enterprises
- Full agent orchestration: multi-step reasoning, tool use, memory management
- Per-user memory: each user has isolated persistent context across sessions
- Multi-model routing: @fastest for quick queries, @best for complex tasks
- Automatic failover: if a model provider degrades, traffic shifts in real time
- Cost control: Rust workers optimize data before inference, cutting token spend
- Compliance: E2EE on every payload, multi-tenant isolation, audit trail via event sourcing
One platform to run your entire AI agent infrastructure. Not five tools duct-taped together.

Roadmap

Solid foundation. Ambitious future.

Shipped
Production-grade orchestration engine
Workflow DSL parser and executor
400+ integrations across 60+ categories
Secure gateway with client auth and encryption
Monitoring dashboard
Multi-tenant isolation
Coming next
Autonomous workflow optimization
Independent per-node autoscaling
High-performance data processing layer
Visual workflow editor
Integration marketplace
ionun CLI

Ready in minutes.

Execute workflows via REST API. Use named workflows or send raw Mermaid directly.

curl -X POST https://gateway.ionun.com/api/workflows/execute \
  -H "Authorization: Bearer <token>" \
  -H "Content-Type: application/json" \
  -d '{
    "workflow_name": "textcorrection",
    "input": {
      "text_content": "Fix ths sentance plz"
    }
  }'