Archi

Archi (AI Augmented Research Chat Intelligence) is a retrieval-augmented generation (RAG) framework designed to be a low-barrier, open-source, private, and customizable AI solution for research and educational support.

Archi makes it easy to deploy AI assistants with a suite of tools that connect to

  • knowledge bases such as web links, files, JIRA tickets, and documentation
  • communication platforms such as Piazza, Slack, Mattermost, and email

It is modular and extensible, allowing users to add connectors and customize pipeline behavior for a wide range of tasks—from answering simple questions to delivering detailed explanations.

Start Here

If you are new to Archi, follow this path:

  1. Install
  2. Quickstart
  3. User Guide

About

Archi is developed by Prof. Paus (MIT Physics), Prof. Kraska (MIT EECS), and their students. It has already been successfully deployed as a user chatbot and technical assistant at SubMIT (the MIT Physics Department's computing cluster) and as an educational assistant for several MIT courses, including 8.01 and 8.511.

What sets Archi apart is that it is fully open source, configurable across foundational models and LLM libraries, and designed for private deployment. Under the hood, Archi is a highly configurable RAG system tailored for educational and scientific support. Given its success, the scope now spans additional MIT classes, CERN, Harvard, and internal deployments such as CSAIL's support staff.

Research Resource Support

Archi serves technical support teams and end users. At SubMIT, it functions both as a user-facing chatbot and as a ticket assistant. Integration with Redmine enables Archi to prepare draft responses to support tickets that staff can review before sending. In both roles, Archi accesses the corpus of tickets and documentation, citing relevant sources in its answers.

Educational Support

Archi assists TAs, lecturers, and support staff—or students directly—by preparing answers based on curated class resources. In MIT course deployments, Archi leverages Piazza posts, documentation, and other class-specific materials. The Piazza integration can draft answers for staff to review or send, while the system continuously learns from revisions and new posts, improving over time.


Documentation

Section Description
Install System requirements and installation
Quickstart Deploy your first instance in minutes
User Guide Overview of all capabilities
Data Sources Configure web links, git, JIRA, Redmine, and more
Services Chat, uploader, data manager, Piazza, Mattermost, and other interfaces
Models & Providers LLM providers (OpenAI, Anthropic, Gemini, OpenRouter, Local), embeddings, BYOK
Agents & Tools Agent specs, tools, MCP integration, pipelines
Configuration Full YAML config schema reference
CLI Reference All CLI commands and options
API Reference REST API endpoints
Benchmarking Evaluate retrieval and response quality
Developer Guide Architecture, contributing, extending the stack
Advanced Setup GPU setup, multi-node, production deployment
Troubleshooting Common issues and fixes