Documentation

Everything you need to get started with Code Assistant

Overview

Code Assistant is a local AI-powered tool that lets you chat with your codebase using RAG (Retrieval-Augmented Generation). Index your code, ask questions, and get intelligent answers with relevant context.

100% Private

Your code stays on your machine. Use local LLMs for complete privacy.

Multiple Providers

Connect to Gemini, OpenAI, OpenRouter, Z.AI, LM Studio, or Ollama.

E2E Encrypted Sync

Optional cloud sync with military-grade encryption. Even we can't read your data.

Getting Started

Get up and running in three simple steps.

1

Create a Project

Go to Projects page, click "New Project", and give it a name. Each project has its own indexed files, chat history, and settings.

2

Configure Your LLM

In Project Settings, choose your AI provider (local or cloud), select a model, and save. For cloud providers, add your API key in Global Settings first.

3

Upload & Index Your Code

Go to the Ingest tab, drag and drop your files or folders, then click "Start Indexing". Once complete, start chatting!

Installation

Code Assistant runs as a desktop application with bundled Python runtime. No additional dependencies required.

System Requirements

Minimum:

  • 8 GB RAM
  • 2 GB disk space
  • Windows 10, macOS 11+, or Linux

Recommended:

  • 16 GB RAM
  • GPU for local LLMs (optional)
  • SSD for faster indexing

Windows

Download the installer (.exe) from GitHub Releases and run it. Python is bundled - no separate installation needed.

macOS & Linux

Download the appropriate package from GitHub Releases. Docker deployment is also available for server environments.

LLM Providers

Choose the AI provider that fits your needs - from free local options to powerful cloud APIs.

Local Providers (Free & Private)

LM Studio

User-friendly GUI for running local models. Great for beginners.

  • Download models with one click
  • Built-in local server
  • No API key required

Ollama

CLI-based tool for power users. Fast model switching.

  • Command-line interface
  • Easy model management
  • No API key required

Cloud Providers

Gemini

Google AI with generous free tier.

Free tier available

OpenAI

GPT-4o and other premium models.

Paid

Z.AI

GLM-4.7 and other GLM models.

Free & paid models

OpenRouter

Access 100+ models from one API.

Free & paid models

Usage Guide

Get the most out of your code conversations.

Example Questions

"What does this project do?"

"Explain the authentication flow"

"Where is the database connection handled?"

"Why might this function return null?"

Tips for Better Results

  • Be specific - "How does the login function validate passwords?" beats "How does login work?"
  • Index relevant files - Only index the code you want to query for faster, more accurate results.
  • Try different models - Some models handle code better than others. Experiment to find your best fit.

Security & Privacy

Your code security is our top priority. Choose your privacy level.

Maximum Privacy (Local)

Use LM Studio or Ollama - your code never leaves your machine.

  • All processing happens locally
  • No internet required after setup
  • Zero data transmission

Cloud with E2E Encryption

For cloud sync users, all data is encrypted before leaving your device.

  • XChaCha20-Poly1305 encryption
  • Argon2id key derivation
  • 24-word recovery key backup

Data Handling by Provider Type

ProviderCode Destination
LM Studio / OllamaStays on your machine
OpenAI / Gemini / Z.AI / OpenRouterSent to APIs (encrypted in transit)

Supported File Types

Code

.py, .js, .ts, .tsx, .jsx, .java, .cpp, .go, .rs, .rb

Web & Config

.html, .css, .json, .yaml, .toml, .env

Documentation

.md, .txt, .rst

Excluded

Binary files, images, videos are automatically skipped

API Key Setup

Troubleshooting

Chat Not Responding

  1. Verify your LLM provider is running (for local providers)
  2. Ensure a model is selected in Project Settings
  3. Check API key validity for cloud providers
  4. Try refreshing the page or restarting the app

Search Results Irrelevant

  1. Be more specific in your questions
  2. Verify files are indexed (check Ingest tab)
  3. Increase retrieval count in Settings
  4. Re-index if you've made major code changes

Embedding Model Mismatch Error

This occurs when trying to index with a different embedding model than originally used. Go to Ingest tab, clear indexed files, then re-index with your preferred model.

Frequently Asked Questions

Is my code sent to the cloud?

It depends on your provider choice. Local providers (LM Studio, Ollama) keep everything on your machine. Cloud providers send code to their APIs, encrypted in transit.

Which provider should I use?

For maximum privacy, use LM Studio or Ollama. For free cloud access, try Gemini. For best quality, OpenAI GPT-4o or Z.AI GLM-4.7. For variety, OpenRouter offers 100+ models.

Can I change the embedding model after indexing?

No, the embedding model is locked per project. To change it, you must clear all indexed files and re-ingest from scratch.

Do I need a GPU?

Not required. Cloud providers work without a GPU. For local LLMs, a GPU helps but CPU-only mode works for smaller models.

Ready to Get Started?

Download Code Assistant and start chatting with your codebase in minutes.