Download and Installation

Download and Installation overview

Before installing Sapientia, please ensure your hardware and operating system meet the minimum requirements to guarantee optimal AI model performance.


System Requirements

Running Large Language Model (LLM) inference typically requires significant memory (RAM).

However, Sapientia is engineered to be a lightweight and accessible for running LLMs. By default, it utilizes the efficient Gemma 3n E2B and EmbeddingGemma 300m models. We extend our sincere appreciation to Google DeepMind, as the developers of the Gemma models, for their outstanding optimization of the Gemma models, which allows Sapientia to run smoothly on consumer hardware—from standard laptops to mobile devices.

Consequently, the minimum specifications required are remarkably low.

Hardware Specifications

ComponentMinimum Specifications
CPUIntel Core i5 Gen 8 or equivalent
RAM8 GB
Storage10 GB available space (SSD)

Note: These specifications reflect our internal testing on an Intel Core i5 Gen 8 device with 8 GB of RAM.

OS Compatibility

Sapientia supports modern 64-bit (x64) desktop operating systems.

  • Windows: Windows 10 (Build 1903+) and Windows 11.
  • Linux: Ubuntu 20.04+, Fedora, and other Debian-based distributions.
  • macOS: Coming Soon.

Download

Select the appropriate installation for your operating system.


Model Setup

Sapientia utilizes a local-first AI architecture, meaning the intelligence runs directly on your device.

Base AI Model and Embedding Model

Sapientia relies on two core components to function:

Base AI Model: This is the primary engine responsible for text generation, reasoning, and understanding user instructions.

Embedding Model: This model converts text into numerical vectors, enabling semantic search capabilities. It allows the AI to "understand" and retrieve context from your local documents (Retrieval-Augmented Generation).

Both models utilize the GGUF (GPT-Generated Unified Format). GGUF is a high-performance binary format optimized for fast loading and efficient inference on CPUs and Apple Silicon, significantly reducing memory overhead compared to traditional formats.

Default Models

Upon the initial initialization of the Sapientia application, it is recommended to download the following optimized default models:

Base Model: Gemma 3n 2B

Embedding Model: EmbeddingGemma

Why these models?

We have selected the Gemma 3n 2B and EmbeddingGemma family by Google DeepMind as the default standard because they offer an exceptional balance between performance and efficiency. They deliver state-of-the-art reasoning and retrieval capabilities while remaining lightweight enough to run smoothly on standard consumer devices (e.g., laptops with 8GB RAM) without requiring dedicated high-end GPUs.


Post-Installation Steps

After successfully installing the application, the next step is to configure the Runtime Environment. Please proceed to the Setup Runtime page.