RESTai

AIaaS (AI as a Service) — Create AI projects and consume them via a simple REST API. RAG, Agents, Visual Logic, Inference — all in one platform.

Get Started Explore Features
RESTai Dashboard
Why RESTai?

Everything you need to deploy AI

A complete AI platform — not just another wrapper. Full Web UI, analytics, security, and enterprise features out of the box.

Multi-Project Platform

RAG with SQL-to-NL and auto-sync, Agents with MCP tools, Block visual logic, and direct Inference — all project types in one platform.

📈

Full Web UI + Analytics

React dashboard with token tracking, cost analytics, latency monitoring, and per-project usage charts. Not just an API.

🤖

Any LLM Provider

OpenAI, Anthropic, Ollama, Gemini, Groq, LiteLLM, vLLM, Azure, and any OpenAI-compatible endpoint.

🔒

Enterprise Security

Teams, RBAC, OAuth/LDAP/OIDC, TOTP 2FA, input/output guardrails, audit logging, and per-project rate limiting.

🔌

Extensible Tools (MCP)

Model Context Protocol support for unlimited agent integrations. Connect any MCP server via HTTP/SSE or stdio.

🎨

White-Label Branding

Custom logos, colors, and app names per team. Built-in knowledge sync from S3, Confluence, SharePoint, and Google Drive.

Analytics

Dashboard & Analytics

Track token usage, costs, latency, and project activity from a centralized dashboard. Daily charts for tokens, costs, and response latency per project — identify performance regressions at a glance.

RESTai Dashboard
Knowledge

RAG — Retrieval-Augmented Generation

Upload documents and query them with LLM-powered retrieval. Multiple vector stores, ColBERT and LLM-based reranking, and natural language to SQL.

  • ChromaDB, PGVector, Weaviate, Pinecone
  • Auto-sync from S3, Confluence, SharePoint, Google Drive
  • Smart chunking analytics with recommendations
  • Built-in evaluation framework with DeepEval
RESTai RAG
RESTai Analytics
Agents

Agents + MCP

Zero-shot ReAct agents with built-in tools and MCP (Model Context Protocol) server support. Connect any MCP-compatible server via HTTP/SSE or stdio for unlimited tool access.

RESTai Agent
Playground

Projects & Chat Playground

Create and manage AI projects with their own LLM, system prompt, tools, and configuration. Test instantly in the built-in chat playground with streaming responses and multimodal support.

RESTai Playground
Visual Logic

Block — Visual Logic Builder

Build processing logic visually using a Blockly-based IDE — no LLM required. Drag-and-drop blocks to define how input is transformed into output. Use the "Call Project" block to compose AI pipelines.

RESTai Block IDE
Quality

Evaluation Framework

Built-in evaluation system to measure and track AI project quality over time. Create test datasets, run evaluations with multiple metrics, and visualize score trends.

  • Answer Relevancy — is the answer relevant?
  • Faithfulness — is it grounded in context? (RAG)
  • Correctness — does it match expected output?
RESTai Evaluation
Versioning

Prompt Versioning

Every system prompt change is automatically versioned. Browse history, compare versions, and restore any previous prompt. Eval runs link to prompt versions for A/B comparison.

RESTai Prompt Versioning
GPU

Image Generation

Local and remote image generators loaded dynamically. Supports Stable Diffusion, Flux, DALL-E, RMBG2, and more. Auto-detects NVIDIA GPUs with detailed hardware monitoring.

RESTai Image Generation
Enterprise

Teams, Branding & Security

Multi-tenant with teams, RBAC, custom branding per team (white-labeling), TOTP 2FA, input/output guardrails, and a full audit log.

  • Custom logos, colors, and app names per team
  • OAuth, LDAP, OIDC single sign-on
  • TOTP 2FA with admin enforcement
  • Full audit trail for all mutations
RESTai Branding
RESTai Audit
Auto-Sync

Knowledge Base Sync

Automatically keep your RAG knowledge base up-to-date by syncing from external sources on a schedule. Configure per project with independent settings.

  • Web URLs — re-scrape pages automatically
  • Amazon S3 — sync documents from buckets
  • Confluence — sync space pages
  • SharePoint — sync via Microsoft Graph
  • Google Drive — sync via service account
RESTai Teams
OpenAI-Compatible

Direct Access API

Use LLMs, image generators, and audio transcription directly via OpenAI-compatible endpoints — no project required. Team-level permissions control access, and all usage counts toward budgets.

  • POST /v1/chat/completions — Chat with any LLM (streaming supported)
  • POST /v1/images/generations — Generate images via DALL-E, Flux, SD, etc.
  • POST /v1/audio/transcriptions — Transcribe audio files

Works with any OpenAI-compatible SDK — just point base_url to your RESTai instance.

RESTai Direct Access
Providers

Any LLM Provider

Any LLM supported by LlamaIndex. Configurable context windows with automatic chat memory management.

Ollama
chat / vision
OpenAI
chat
Anthropic
chat
Google Gemini
chat / vision
Groq
chat
Grok (xAI)
chat
LiteLLM
chat
vLLM
chat
Azure OpenAI
chat
OpenAI-Compatible
chat
Quick Start

Up and running in 30 seconds

Install from PyPI
# Install
pip install restai-core

# Setup database
restai init
restai migrate

# Start server
restai serve

# Open http://localhost:9000/admin
# Login: admin / admin
From Source (Development)
# Clone and install
git clone https://github.com/apocas/restai
cd restai && make install

# Start dev server
make dev

# Open http://localhost:9000/admin
# Login: admin / admin

With env file: restai serve -e .env -p 8080 -w 4  ·  Docker: docker compose --env-file .env up --build