The Full-Stack AI Operations Platform
PlatformHero unifies multi-model inference, intelligent assistants, visual agent workflows, and RAG knowledge into a single platform — available as a fully managed SaaS, deployable as your Private AI on-premise, or hosted in the cloud.
Multi-Model AI Inference
Connect 35+ LLM providers including OpenAI, Anthropic Claude, Google Gemini, Azure OpenAI, AWS Bedrock, DeepSeek, Mistral, Cohere, Groq, xAI, Ollama, and LM Studio. Support for chat completion (streaming + function calling), text embedding, and rerank. Automatic model fallback ensures uninterrupted service.
AI Assistants
Build configurable AI assistants with model binding, system prompts with variables, memory modes, tool integration, and RAG-powered knowledge. Manage stateful chat sessions with streaming, debug mode, and granular per-user access control. Track per-assistant token usage, cost, and session statistics with daily breakdowns.
Agent Workflows (DAG Engine)
Design complex automation with a visual directed acyclic graph editor. 25+ node types: triggers (manual, webhook, cron), AI nodes (LLM chat, embedding), tool nodes (plugin, action, MCP), knowledge retrieval, logic (condition, switch, loop, merge, wait), data (HTTP, code, JSON transform), output, and sub-assistant calls.
Knowledge Base / RAG
Vector-powered knowledge management with Qdrant. Organize data into collections, records, and chunks. 8 content loaders: PDF, DOCX, XLSX, CSV, HTML, Markdown, TXT, and web URL. Let your assistants and agents answer from your proprietary data.
27+ Plugins, Actions & MCP
Built-in plugins for Bing, Google, DuckDuckGo search, web scraping, Wikipedia, GitHub, code interpreter, chart maker, weather, and more. Create custom actions with auth. Connect any MCP server (STDIO/SSE/HTTP) for automatic tool discovery and infinite extensibility.
Chat Interface
Full-featured conversational UI with assistant selector, streaming responses, thinking visualization, multimodal input (text + images + files), artifacts (code/document/sheet/image editors), suggested actions, tool approval, and persistent chat history.
Enterprise Platform Management
Flexible workspace management with plan-based quotas for users, assistants, collections, storage, agents, and daily rate limits. Workspace lifecycle management with expiration and suspension. Built-in secure licensing and activation system.
Auth & Security
Role-Based Access Control with Admin, Builder, and User roles. JWT multi-session token management, AES credential encryption at rest, configurable rate limiting. Email-based flows for signup, verification, password reset, and welcome onboarding.
Logging & Analytics
Complete inference logging with token usage and cost tracking per request. Per-assistant and per-workspace usage dashboards. Storage consumption monitoring. Prometheus metrics and health check endpoints for observability.
Console & Admin Panel
Full management console: Dashboard, Models, Playground (chat + assistant modes), Assistants, Agents (visual editor), Knowledge, Plugins/Actions/MCP, Logs, Users, API Keys, and Settings. Separate Admin panel for workspace management and activation code provisioning.
OpenAI-Compatible API
Drop-in replacement for the OpenAI SDK. Point your existing applications at PlatformHero and use any of the 35+ connected providers through a unified, standard API. Perfect for migration without code changes.
Flexible Deployment
Use the fully managed SaaS to get started instantly, or deploy on-premise as your Private AI — all data and models stay on your infrastructure. Manage rollouts through the deployment manager with web UI.