Messaging protocol daemon for Email (IMAP/SMTP), SMS (Twilio), WhatsApp (Business API), Signal (E2E encrypted), Microsoft Teams (Graph API), and XMPP/Jabber.
Features: AI auto-response with approval workflow, unified message format, semantic search, webhook handling on port 8896
Multi-agent orchestration with tiered swarm hierarchy for autonomous goal execution.
Features: Task decomposition, agent coordination, UAP protocol, MCP support
Binary:eldric-swarmd
Deployment Scenarios
1. Development Setup
Single machine for local development and testing
2. Production Multi-Region
Global deployment with regional controllers
3. Enterprise Mainframe
IBM z/OS DB2 connectivity via DRDA
4. Knowledge Routing
AI-powered theme detection and model selection
5. Swarm Multi-Agent Orchestration
Autonomous goal execution with tiered agent hierarchy
Use Cases
Enterprise AI Analytics Platform
User Query
→
Edge/CDN
→
Router Theme Detection
→
AI Worker SQL Generation
→
Data Worker Query DB
→
AI Worker Analyze Results
Multi-Database RAG Pipeline
Combine data from multiple enterprise databases for AI-powered analysis:
# AI Worker queries multiple data sources1. Query PostgreSQL for customer data
2. Query MySQL for transaction history
3. Query DB2 z/OS for mainframe records
4. AI synthesizes all data into unified response
Real-time Business Intelligence
Dashboard Request
→
Controller API
→
Data Workers
→
DB Cluster
→
AI Summary
Voice-Enabled AI Assistant
End-to-end voice chat using Media Worker for STT/TTS:
Voice Input Audio Stream
→
Media Worker STT (Whisper)
→
AI Worker LLM Response
→
Media Worker TTS (Piper)
→
Voice Output Audio Stream
Video Meeting Summarization
Transcribe and analyze recorded meetings:
# Video processing pipeline1. Media Worker extracts audio from video
2. Media Worker transcribes with speaker diarization
3. Data Worker stores transcript with vector embeddings
4. Agent Worker generates meeting summary and action items
5. Comm Worker distributes summary via email/Teams
Unified Customer Communication
AI-powered multi-channel messaging with Comm Worker:
Customer WhatsApp/Email/SMS
→
Comm Worker Protocol Adapter
→
Data Worker Message RAG
→
AI Worker Draft Response
→
Comm Worker Send Reply
Agentic Knowledge Search
Complex question answering with iterative retrieval:
# Agent Worker performs multi-hop reasoning1. Agent Worker receives complex query
2. Decomposes into sub-questions (Query Decomposition)
3. Retrieves relevant documents from Data Worker (RAG)
4. Iteratively refines search based on findings (ReAct)
5. Synthesizes final answer with citations
Individual agent execution context with tool access, memory management, and inter-agent communication.
Execution Modes
Advisory
Swarm proposes actions, waits for human approval before each step. Maximum oversight.
Supervised
Runs autonomously but requires approval for critical operations (writes, executions, external calls).
Autonomous
Full autonomous operation with real-time monitoring. Best for trusted, well-defined goals.
Port Reference
Component
Default Port
Protocol
Purpose
Controller
8880
HTTP/REST
Cluster API, Dashboard, Node Registration
Router
8881
HTTP/REST
Request Routing, Load Balancing
AI Worker
8890
HTTP/REST
LLM Inference, Tool Execution
Cloud Worker
8889
HTTP/REST
Multi-Backend Cloud Inference Gateway
Data Worker
8892
HTTP/REST
Storage, Database, Vector/RAG
Agent Worker
8893
HTTP/REST
Agentic RAG, Multi-Agent Execution
Media Worker
8894
HTTP/REST
STT, TTS, Video Processing
Comm Worker
8895
HTTP/REST
Email, SMS, WhatsApp, Signal, Teams, XMPP
Comm Webhooks
8896
HTTP
Incoming Message Webhooks
IIoT Worker
8891
HTTP/REST
Industrial IoT, OPC UA, Modbus, MQTT
Science Worker
8897
HTTP/REST
Bioinformatics, Pharma, CRISPR, LIMS
Training Worker
8898
HTTP/REST
AI Model Training, Fine-tuning
Edge Server
443
HTTPS
Unified Proxy, TLS, API Auth, Rate Limiting
Swarm Orchestrator
8885
HTTP/REST
Goal Management, Task Decomposition
Swarm Controller
8886
HTTP/REST
Agent Lifecycle, Task Scheduling
Agent Session
8887
HTTP/REST
Agent Execution, Tool Access
Data NFS
2049
NFS
Filesystem Access via NFS-Ganesha
Reserved Ports (Inference Backends)
Backend
Port
Notes
Ollama
11434
Local LLM runtime
vLLM
8000
OpenAI-compatible API
TGI
8080
HuggingFace Text Generation
llama.cpp
8081
Native GGUF serving
Triton
8000-8002
NVIDIA multi-framework
TensorFlow Serving
8501
TensorFlow models
Model Distribution Flow
Models flow through the cluster via two primary paths: from public registries or from the internal Data Worker registry.
Path 1: Public Registry Distribution
Ollama Hub Public Models
→
Controller Model Registry
→
AI Workers Local Cache
The controller tracks which models are available on which workers. When a model is requested that isn't cached locally, the worker pulls it from Ollama Hub on demand.
Path 2: Data Worker Registry Distribution
Data Worker Model Storage
→
Controller Distribution
→
AI Workers Deploy & Serve
Fine-tuned or custom models stored on Data Workers can be distributed to inference workers via the controller. This enables private model registries without relying on public hubs.
Cloud Worker Model Routing
Client Request model: "gpt-4o"
→
Router Model Lookup
→
Cloud Worker cld- prefix
→
OpenAI / xAI / Anthropic Cloud API
Cloud Workers auto-discover models from connected cloud backends and register them with the controller. The router routes requests to the correct Cloud Worker based on model name.
Path 3: Distributed Pipeline Inference New
Data Worker GGUF Model Store
→
Controller GGUF Parser + Shard Coordinator
→
Workers (N) Each loads assigned layers via NFS
For models too large for a single GPU, the Controller reads the GGUF file metadata from the Data Worker, calculates layer assignments proportional to each worker's available VRAM, and pushes shard configurations to workers. Workers start llama-rpc-server (middle/tail) or llama-server --rpc (head). The router auto-discovers the pipeline model from the head worker's heartbeat.
Sepp Hochreiter's extended LSTM architecture (xLSTM) is integrated across the cluster for predictive workloads, training, and anomaly detection.
Router: xLSTM Predictor
Trained xLSTM model in the router for workload forecasting. Predicts load spikes before they happen, enables pre-emptive scaling. Fast sequence classification for intent detection without LLM overhead. Anomaly detection flags unusual traffic patterns.
Training Worker: xLSTM Backend
Native xLSTM training backend with sLSTM (scalar LSTM with exponential gating) and mLSTM (matrix LSTM with covariance update) cell support. Fine-tune xLSTM models for domain-specific sequence tasks — time series, genomics, financial data.
Science Worker: Anomaly Detection
xLSTM-based anomaly detection on scientific time-series data. Trained on seismic, genomic, financial, and climate datasets. Detects outliers in real-time sensor streams from IoT and LIMS integrations.
Swarm LLM Ensemble
The Router's Swarm LLM feature sends queries to multiple models simultaneously and combines results using consensus strategies for higher-quality responses.
Ensemble Strategies
debate
Models argue over the answer, judge LLM picks the best reasoning
critique
First model generates, second critiques, synthesis from both
best_of_n
N responses generated, judge model scores and selects best
vote
Majority consensus across all participating models
How It Works
The router selects multiple workers with different models (e.g., Llama + Qwen + Mistral), sends the same query to all, then applies the selected ensemble strategy to produce a final response.
Client → Edge → Router (swarm=debate)
→ Worker1 (llama3.1) → Response A
→ Worker2 (qwen2.5) → Response B
→ Worker3 (mistral) → Response C
Router → Judge LLM → Best response → Client
Eldric cluster nodes are location-independent. Spread workers across university labs, corporate datacenters, home offices, and cloud providers. Workers connect through the Edge TLS gateway — no VPN required. Workers behind NAT use the built-in tunnel to receive inference requests through outbound-only connections.
Direct Registration via Edge
Workers with internet access register through the Edge TLS gateway. API key authentication, rate limiting, and encryption are handled by the Edge. The worker just sets --controller https://edge.example.com
Tunnel for NAT/Firewall
Workers behind NAT connect outbound to the Edge tunnel. They long-poll for inference requests — no inbound ports, no VPN, no public IP needed. The Edge queues requests and delivers them through the tunnel.
VPN / Private Network
For sites already connected via VPN or WireGuard, workers register directly with the controller on the private network. No Edge proxy needed — just --controller http://ctrl:8880
# Central datacenter (Vienna) — direct
./eldric-controller --port 8880
./eldric-edge --port 443 --cert /etc/ssl/cert.pem --key /etc/ssl/key.pem --controller http://localhost:8880
# University Lab A (Zurich) — registers through Edge
ssh zurich-lab "./eldric-workerd --backend ollama --controller https://edge.example.com --api-key sk-lab-a"
# University Lab B (Munich) — registers through Edge
ssh munich-lab "./eldric-workerd --backend vllm --controller https://edge.example.com --api-key sk-lab-b"
# Remote researcher (home office, behind NAT) — uses tunnel
./eldric-workerd --backend ollama --tunnel https://edge.example.com --api-key sk-remote-1