An open-source platform that connects to 40+ apps to provide intelligent search and AI assistance across all company information

Overview:

Onyx is an open-source AI platform that provides an application layer for large language models (LLMs). It enables LLMs with advanced capabilities including retrieval-augmented generation (RAG), web search, code execution, file creation, and deep research. Onyx is designed to be self-hosted by individuals or teams, with deployment options ranging from a lightweight Chat UI (Lite mode) for quick testing to a full-featured stack for serious users and larger organizations. It includes over 50 indexing-based connectors for integrating with external applications.

Core Features:

  • Agentic RAG: Hybrid indexing combined with AI agents for information retrieval to deliver search and answer quality.

  • Deep Research: Generates in-depth reports using a multi-step research workflow.

  • Custom Agents: Build AI Agents with unique instructions, knowledge, and actions.

  • Actions & MCP: Allows agents to interact with external applications, with flexible authentication options.

  • Code Execution: Executes code in a sandboxed environment for data analysis, graph rendering, or file modification.

  • Artifacts: Generates downloadable documents, graphics, and other file types.

Use Cases:

  • Developers building AI applications: Integrating LLMs with custom agents, web search, and code execution into existing workflows.

  • Teams deploying collaborative AI chatbots: Sharing chats and agents across an organization with SSO, RBAC, and query history.

  • Self-hosters evaluating LLM capabilities: Testing a lightweight Chat UI (Lite mode) or deploying a full RAG and indexing pipeline locally.

  • Researchers conducting multi-step analysis: Using the Deep Research feature to generate comprehensive reports from aggregated sources.

Why It Matters:

Onyx offers a self-hosted AI platform that supports both proprietary and self-hosted LLM providers, including Ollama, LiteLLM, vLLM, Anthropic, OpenAI, and Gemini. Its two-tier deployment model (Lite vs. Standard) lets users start with minimal resources (under 1GB memory) and scale up to a full RAG stack with vector indexes, job queues, and caching. The MIT-licensed Community Edition covers core Chat, RAG, Agents, and Actions features, while the Enterprise Edition adds SSO, RBAC, analytics, and whitelabeling for larger organizations.

TeilenXLinkedInReddit

Ähnliche Tools

Projektstatistiken

Sterne

28,836

Forks

3,866

Lizenz

Unknown

Metadaten

Alternative zu
Algolia