Every Feature You Need
to Run AI at Scale

Eldric is the distributed agentic AI platform for serious deployments. Built in Austria — the birthplace of LSTM and xLSTM (Sepp Hochreiter). From a single laptop to a 50-node cluster, from local Ollama to cloud APIs — one platform, fully under your control.

13
Specialized Agents
40+
Built-in Tools
13
Worker Types
20+
LLM Backends
140+
Science APIs
5
Swarm Strategies

Distributed Architecture

Multi-tier architecture with specialized workers, intelligent routing, and horizontal scaling.

Clients (Web, API, Mobile) Edge Server (TLS, Auth, Plugins) Router (Swarm LLM, xLSTM) debate | critique | vote | best_of_n | synthesize Controller (Ops, Topology) pipeline sharding | cluster ops Inference Workers Ollama, vLLM, TGI Cloud Workers OpenAI, Anthropic, xAI Data Workers Storage, DB, RAG Agent Workers Agentic RAG, Multi-agent Media Workers STT, TTS, Video Comm Workers Email, SMS, WhatsApp Science Workers 140+ APIs, LIMS Training Workers LoRA, DPO, MLX IoT Workers OPC-UA, Smart Home Swarm Controller Multi-agent orchestration Inference Workers Native GGUF, xLSTM NOVA Self-improving AI All workers auto-register with Controller. Router syncs models, health, and xLSTM predictions. Large models can be sharded across workers via pipeline parallelism (distributed inference).

Swarm LLM New

Multiple LLMs reasoning together as a collective intelligence. Auto-selects the best strategy based on your query.

Debate

Models argue positions across rounds. A judge renders the final verdict. Best for decisions.

🔎

Critique

Model A generates, Model B critiques, A refines. Iterates for quality. Best for writing.

Best of N

Fan-out to N models, judge picks the single best answer. Best for code generation.

Vote

All models answer, consensus analysis with confidence. Best for factual questions.

Synthesize

Merge insights from all models into one comprehensive answer. Default strategy.

Auto-Selection

The router automatically picks the optimal strategy based on query content: “Kafka vs RabbitMQ?” triggers debate, “Write a migration plan” triggers critique, “Is this thread-safe?” triggers vote. Override with ensemble_strategy in the API.


13 Specialized Agents

Each agent constrains available tools for focused, safe execution.

GeneralAll tools
ExplorerRead-only navigation
CoderRead, Write, Edit
RunnerBash, Read
PlannerArchitecture & design
SearcherWeb research
DatabaseSQL operations
LearnerRAG / knowledge
NetworkSSH, DevOps
TrainerModel fine-tuning
MergerModel combination
OrchestratorCluster management
ResearcherDeep research + sources

Platform Features

Everything you need to deploy, scale, and operate AI infrastructure.

20+ LLM Backends

Ollama, vLLM, TGI, llama.cpp, MLX, OpenAI, Anthropic, Azure, Groq, xAI, Together, Mistral, Cohere, NVIDIA Triton, TensorFlow Serving, TorchServe, ONNX Runtime, and more. Switch backends without changing code.

Free
📊

AI-Powered Load Balancing

Intelligent routing based on model specialization, load metrics, latency, and query intent. The router uses an LLM for real-time decisions, with optional xLSTM predictor for workload forecasting, anomaly detection, and fast classification.

Enterprise
🔍

RAG & Vector Storage

Built-in vector database with SQLite, FAISS, or ChromaDB backends. Automatic text chunking, embedding generation (Ollama, OpenAI), and semantic search. Multi-tenant with namespace isolation.

Standard
🚀

Edge Server & Web Client

TLS termination, API key auth, rate limiting, embedded web chat UI with PWA support. Proxies worker registration for geo-distributed clusters. Built-in NAT tunnel for workers behind firewalls — no VPN needed.

Free
🌎

Worldwide GPU Clusters New

Spread workers across university labs, corporate datacenters, home offices, and cloud providers. Workers register through the Edge TLS gateway over the internet. Mix RTX 3090s, H100s, and Apple Silicon in the same cluster. Architecture →

Standard
🔒

NAT Tunnel New

Workers behind NAT/firewalls receive inference requests through outbound-only long-poll connections to the Edge. No inbound ports, no VPN, no public IP. Just --tunnel https://edge.example.com. Edge docs →

Free
🤖

Multi-Agent Swarm

Autonomous goal decomposition, multi-agent collaboration, agent marketplace, human-in-the-loop approval, and visual swarm dashboard with real-time topology.

Professional
🎓

Model Training & Fine-tuning

LoRA, QLoRA, SFT, DPO, RLHF via Unsloth, Axolotl, TRL, DeepSpeed, MLX. Native xLSTM training backend for Hochreiter's extended LSTM architecture. Training chains with visual node editor. Latent reasoning: COCONUT, Quiet-STaR, Pause Tokens, Hidden CoT, DeepSeek DSA.

Standard
🎙

Media Processing

Speech-to-text (Whisper.cpp, OpenAI), text-to-speech (Piper, ElevenLabs), voice chat, video transcription, scene detection, multimedia RAG indexing.

Free (basic)
💬

Communication Protocols

Email (IMAP/SMTP), SMS (Twilio), WhatsApp (Business API), Signal (E2E encrypted), Microsoft Teams, XMPP. AI auto-response with approval workflow.

Professional
🔬

140+ Science APIs

NASA, ESA, CERN, LIGO, JWST, PubMed, ClinicalTrials.gov, NCBI, UniProt, Materials Project, GBIF. Bioinformatics, CRISPR, pharmaceutical, LIMS with GLP/FDA compliance. xLSTM-powered anomaly detection on scientific datasets.

Free (basic)
🔌

Plugin System New

Extensible plugin architecture: tools, filters, widgets, integrations. Public catalog at eldric.ai/plugins with one-click install. Per-user preferences synced to server.

Free
💻

Cluster Operations New

Remote command execution, script deployment, log collection, rolling restarts, and system status across all cluster nodes via the Controller API.

Enterprise

Distributed LLM Inference New

Split models larger than any single GPU across multiple workers. GGUF layer-level sharding with VRAM-aware scheduling. Automated llama.cpp RPC orchestration — deploy a 70B model across 3 workers with one API call. Learn more →

Standard

Native Inference Worker New

Direct GGUF model loading via embedded llama.cpp — no Ollama, no vLLM, no external dependencies. Permanent VRAM residence for zero cold-start latency. CUDA + Metal GPU support, continuous batching, xLSTM models, pipeline parallelism, and Data Worker model distribution. Learn more →

Free
🔒

Enterprise Licensing

Ed25519-signed licenses with hardware binding. Online/offline validation. Per-backend worker limits. Five tiers: Free, Standard, Professional, Enterprise, Custom.

Free tier available
🌐

Native Tool Calling

Automatic detection: native tool calling for supported models (Ollama 0.3+, OpenAI, Anthropic), XML fallback for others. 40+ built-in tools across all categories.

Free
📱

Multi-Platform Clients

C++ CLI for terminal, native macOS SwiftUI GUI, iOS app, embedded web chat, OpenWebUI compatible. All clients share the same distributed backend.

Free
🛠

Agent Builder & Generator

Agent Builder: LLM-powered creative agent design. Agent Generator: template-based mass production. Both deploy directly to your cluster.

Standard

xLSTM Integration New

Native support for Sepp Hochreiter's xLSTM architecture. Dedicated xLSTM training backend, xLSTM predictor in the router for workload forecasting and fast sequence classification, and xLSTM-based anomaly detection across workers.

Standard
🛰

Industrial IoT (OPC-UA)

Smart home integration, Industrial IoT with OPC-UA protocol, real-time sensor monitoring, AI-powered anomaly detection on production data.

Professional

License Comparison

Choose the tier that fits your deployment. Free tier includes everything you need to get started.

FeatureFreeStandardProfessionalEnterprise
Infrastructure
Controllers1125+
Routers12410+
Inference Workers (Proxy)231050+
Native Inference Workers1310Unlimited
Edge Servers125Unlimited
AI & Inference
LLM BackendsAll 20+AllAllAll
Streaming SSE
Native Tool Calling
Swarm LLM (ensemble)
AI-Powered Routing
Data & Knowledge
RAG / Vector Storage
Embeddings
Database ConnectivitySQLite+PG, MySQL+DB2All
NFS Storage
Agents & Orchestration
13 Agent Types
Multi-Agent (parallel)
Swarm Orchestration
Agent Builder3/month20/month100/monthUnlimited
Media & Communication
STT (Speech-to-Text)Basic+Streaming+DiarizationAll
TTS (Text-to-Speech)Basic+Streaming+Voice CloneAll
Comm ProtocolsEmail+SMSAll except SignalAll (incl. E2E)
Science & Training
Science APIs (140+)Basic+BLAST+Docking, ADMETAll + AlphaFold
Training BackendsMLX only+Unsloth, TRL, xLSTMAllAll + Distributed
xLSTM Predictor (Router)
Training Chains
LIMS / 21 CFR Part 11GLPFull FDA
Operations & Security
Plugin System
Cluster Ops (remote exec)
Dashboard & Metrics
Multi-Tenant Isolation
Hardware Binding
Priority Support

Supported Backends

Connect to any LLM runtime — local or cloud. All backends support streaming.

BackendStreamingVisionNative ToolsEmbeddingsType
OllamaLocal
vLLMLocal
llama.cppLocal
MLXLocal (Apple)
TGILocal
NVIDIA TritonInference
OpenAICloud
AnthropicCloud
xAI (Grok)Cloud
GroqCloud
Together AICloud
HuggingFaceCloud

Ready to deploy?

Download Eldric and have your AI infrastructure running in minutes.

Download Eldric View Licensing