
Boomerang for Gmail
One click calendar scheduling plus powerful email management tools.

The privacy-first, open-source ChatGPT alternative that runs 100% offline.

Jan is an industry-leading open-source AI orchestration platform designed to run LLMs entirely on-device, ensuring zero data leakage to external servers. Built on a modular C++ architecture known as Nitro, Jan optimizes inference across diverse hardware architectures, including NVIDIA GPUs (via TensorRT and CUDA), Apple Silicon (via Metal), and generic CPUs. By 2026, Jan has positioned itself as the definitive desktop 'AI Operating System' for developers and enterprises who require high-performance local inference without the vendor lock-in or privacy risks of cloud-based models. Its core value proposition lies in its OpenAI-compatible local server, which allows it to act as a drop-in replacement for proprietary APIs in existing development workflows. The platform features a robust extension framework that supports local Retrieval-Augmented Generation (RAG), multimodal vision capabilities, and integration with local vector databases. Its commitment to transparency and data sovereignty makes it a critical tool for regulated industries, including healthcare, legal, and government sectors, where data privacy is non-negotiable.
Jan is an industry-leading open-source AI orchestration platform designed to run LLMs entirely on-device, ensuring zero data leakage to external servers.
Explore all tools that specialize in document q&a (rag). This domain focus ensures Jan delivers optimized results for this specific requirement.
A lightweight, modular C++ inference engine that powers model execution with minimal overhead.
Native support for NVIDIA's TensorRT for hyper-optimized inference on RTX GPUs.
Exposes a local server that mimics the OpenAI API structure (v1/chat/completions).
A filesystem-based plugin architecture allowing for third-party tools and UI enhancements.
Built-in vector store capabilities to index and query local documents offline.
Granular control over YAML-based model configurations including temperature, top-p, and frequency penalty.
Unified interface for GGUF, TensorRT, and ONNX model formats.
Navigate to the official Jan.ai website or GitHub repository release page.
Download the installer corresponding to your operating system (Windows, macOS, or Linux).
Execute the installer and complete the standard installation wizard.
Launch Jan and navigate to the 'Hub' or 'Models' tab to view available open-source models.
Select a model (e.g., Llama 3.1, Mistral, or Phi-3) and click 'Download'.
Go to 'Settings' and select 'Advanced Settings' to configure hardware acceleration (CUDA for NVIDIA or Metal for Mac).
Adjust the 'Threads' and 'Context Length' parameters based on your system RAM and VRAM availability.
Initiate a new chat thread to test local inference performance.
Navigate to the 'Local Server' tab to enable the OpenAI-compatible API endpoint (defaulting to port 1337).
Optionally, install the 'RAG' extension to begin indexing local folders for private document interrogation.
All Set
Ready to go
Verified feedback from other users.
"Users praise Jan for its clean UI and superior performance on Apple Silicon compared to competitors. The local API server is frequently cited as a 'killer feature' for developers."
Post questions, share tips, and help other users.

One click calendar scheduling plus powerful email management tools.

The Ultimate AI-Powered Virtual Assistant for Windows and OS Automation

A virtual PDF printer that allows you to create PDF documents from any Microsoft Windows application.

Automate meeting scheduling and calendar synchronization across multiple platforms.

A modular, open-source office and creative suite built for high-performance productivity.

AI-powered email client for smarter, faster, and more secure email management.

A studio for your mind: Object-based note-taking for structured thinking.

Professional-grade browser intelligence and document synthesis agent for research-intensive workflows.