███
░░░
█████ █████ ████ ████████ █████ ████ ████████ ██████
░░███ ░░███ ░░███ ░░███░░███░░███ ░███ ░░███░░███ ███░░███
░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███████
░░███ ███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███░░░
░░█████ █████ ░███████ ░░████████ ████ █████░░██████
░░░░░ ░░░░░ ░███░░░ ░░░░░░░░ ░░░░ ░░░░░ ░░░░░░
░███
█████
░░░░░
A minimal memory layer for AI agents.
In Finnish mythology, Antero Vipunen is a giant who sleeps underground, holding all the world's knowledge and ancient songs. vipune is your agent's sleeping giant — a local knowledge store that remembers everything.
Store semantic memories, search by meaning, and detect conflicts. Single binary CLI. No API keys required.
- Semantic search - Find memories by meaning, not keywords (ONNX embeddings, bge-small-en-v1.5)
- Conflict detection - Automatically warns when adding duplicate or similar memories
- Zero configuration - Works out of the box (auto-detected git projects, sensible defaults)
- Single binary - Just one CLI tool, no daemon, no database server
- No API keys - Everything runs locally, no network dependencies
- Project scoped - Memories isolated by git repository
Supported: macOS ARM64, Linux x86_64, Linux ARM64
Not supported: Windows (due to ONNX Runtime compilation complexity)
curl --proto '=https' --tlsv1.2 -LsSf https://github.com/randomm/vipune/releases/latest/download/vipune-installer.sh | shThe installer detects your platform, downloads the correct binary, and adds it to
~/.cargo/bin/by default. If you don't use Rust, restart your shell or runexport PATH="$HOME/.cargo/bin:$PATH"to make vipune available. Use--prefix=/usr/localfor system-wide installs.
To verify the download before running, download the checksum file and check:
curl -sSfLO https://github.com/randomm/vipune/releases/latest/download/vipune-installer.sh.sha256
sha256sum -c vipune-installer.sh.sha256
sh vipune-installer.shmacOS Apple Silicon (arm64)
Download and extract:
curl -sSfLO https://github.com/randomm/vipune/releases/latest/download/vipune-aarch64-apple-darwin.tar.xz
tar xf vipune-aarch64-apple-darwin.tar.xz --strip-components=1
sudo mkdir -p /usr/local/bin && sudo mv vipune /usr/local/bin/Linux x86_64
Download and extract:
curl -sSfLO https://github.com/randomm/vipune/releases/latest/download/vipune-x86_64-unknown-linux-gnu.tar.xz
tar xf vipune-x86_64-unknown-linux-gnu.tar.xz --strip-components=1
sudo mkdir -p /usr/local/bin && sudo mv vipune /usr/local/bin/Linux ARM64
Download and extract:
curl -sSfLO https://github.com/randomm/vipune/releases/latest/download/vipune-aarch64-unknown-linux-gnu.tar.xz
tar xf vipune-aarch64-unknown-linux-gnu.tar.xz --strip-components=1
sudo mkdir -p /usr/local/bin && sudo mv vipune /usr/local/bin/Requires the Rust toolchain (1.85+). On Linux, you may also need libgomp1 and libc6.
Latest release (recommended)
cargo install vipuneOr clone and build manually
git clone https://github.com/randomm/vipune.git
cd vipune && cargo build --releaseThe binary will be at ./target/release/vipune. Install it:
sudo mkdir -p /usr/local/bin && sudo cp target/release/vipune /usr/local/bin/Or add to PATH temporarily:
export PATH="$(pwd)/target/release:$PATH"rm ~/.cargo/bin/vipune # or: sudo rm /usr/local/bin/vipuneOr via cargo: cargo uninstall vipune
Optionally, clear all data:
rm -rf ~/.vipune ~/.config/vipunevipune downloads the embedding model from HuggingFace Hub on first run. The model is pinned to a specific revision to ensure reproducibility:
- Model ID:
BAAI/bge-small-en-v1.5 - Revision:
main
For air-gapped environments, pre-fetch the model before going offline:
# Install huggingface-cli first if you don't have it
pip install huggingface_hub
# Download the model revision to the cache directory
huggingface-cli download BAAI/bge-small-en-v1.5 \
--revision main \
--cache-dir ~/.cache/huggingface/hubThe model will be cached in the HF Hub cache layout at ~/.cache/huggingface/hub/ and vipune will use it without network access. You can verify the cache before going offline:
ls ~/.cache/huggingface/hub/models--BAAI--bge-small-en-v1.5/Note: When upgrading vipune, the pinned revision may change. Check the release notes and re-download the new revision if the SHA has changed.
Library consumers: The constants vipune::EMBED_MODEL_ID and vipune::EMBED_MODEL_REVISION are exported for tracking the pinned model version programmatically.
Add a memory:
vipune add "Alice works at Microsoft"Search by semantic meaning:
vipune search "where does alice work"Add with metadata (optional):
vipune add "Auth uses JWT tokens" --metadata '{"topic": "authentication"}'| Command | Description |
|---|---|
vipune add <text> |
Store a memory (with type, status, conflict detection) |
vipune search <query> |
Find memories by meaning (filters by type/status) |
vipune get <id> |
Retrieve a memory by ID |
vipune list |
List memories (default: active only) |
vipune delete <id> |
Delete a memory |
vipune update <id> [text] |
Update content, metadata, type, or status |
vipune validate <text> |
Check if text is within embedding token limits |
vipune version |
Show version |
Complete CLI reference • Quickstart guide • Search guide • Architecture
vipune can also be used as a Rust crate for programmatic integration:
# Cargo.toml
[dependencies]
vipune = "0.3.0"use vipune::{Config, MemoryStore, detect_project};
// Initialize memory store
let config = Config::default();
let mut store = MemoryStore::new(
config.database_path.as_path(),
&config.embedding_model,
config.clone()
).expect("Failed to initialize store");
// Add a memory
let project_id = "my-project";
let memory_id = store.add(&project_id, "Alice works at Microsoft", None)
.expect("Failed to add memory");
// Search memories
let results = store.search(&project_id, "where does alice work", 10, 0.0, None, None)
.expect("Failed to search");
for memory in results {
println!("{:.2}: {}", memory.similarity.unwrap_or(0.0), memory.content);
}v0.3 features: MemoryType, MemoryStatus, supersedes flag, and filter parameters are available for type-aware memory management. See the CLI reference for details.
See the crate documentation at docs.rs for complete API reference.
vipune works with zero configuration. All paths use the user's home directory:
Default paths:
- Database:
~/.vipune/memories.db - Model cache:
~/.vipune/models/ - Config file:
~/.config/vipune/config.toml
Environment variables (override defaults):
VIPUNE_DATABASE_PATH- SQLite database locationVIPUNE_EMBEDDING_MODEL- HuggingFace model ID (default:BAAI/bge-small-en-v1.5)VIPUNE_MODEL_CACHE- Model download cache directoryVIPUNE_PROJECT- Project identifier (overrides auto-detection)VIPUNE_SIMILARITY_THRESHOLD- Conflict detection threshold, 0.0-1.0 (default:0.85)VIPUNE_RECENCY_WEIGHT- Recency bias in search results, 0.0-1.0 (default:0.3)VIPUNE_HYBRID- Enable hybrid search (semantic + BM25), true/false or 1/0
Config file (~/.config/vipune/config.toml):
database_path = "~/.vipune/memories.db"
embedding_model = "BAAI/bge-small-en-v1.5"
model_cache = "~/.vipune/models"
similarity_threshold = 0.85
recency_weight = 0.3vipune can act as an MCP (Model Context Protocol) server, enabling AI agents like Claude Code and Cursor to use it as a native memory provider.
Note: MCP is enabled by default. Library users can use
default-features = falsefor sync-only builds.
Add to your Claude Code configuration (~/.claude/settings.json or project .claude.json):
{
"mcpServers": {
"vipune": {
"command": "vipune",
"args": ["mcp"]
}
}
}Add to your Cursor MCP configuration with the same JSON structure.
- store_memory: Store information for later recall
- search_memories: Find memories by meaning (supports
hybridparam) - list_memories: List recent memories
- supersede_memory: Replace an existing memory with new content
vipune works with any agent that can run shell commands — no plugins, adapters, or API keys required. Configure your agent with a few lines of instructions, grant shell command permissions, and the agent can use vipune search and vipune add to maintain persistent memory across tasks.
→ See Agent Integration Guide for per-tool setup instructions (Claude Code, Cursor, Windsurf, Cline, Roo Code, GitHub Copilot, Goose, Aider, OpenCode, Zed, and more).
See also: Search Guide for agent-friendly query patterns.
Exit codes for agent workflows:
0- Success1- Error (missing file, invalid input, etc.)2- Conflicts detected (similar memories found)3- Content too long (exceeds embedding token limit)
Search results can be weighted by recency using the --recency flag or VIPUNE_RECENCY_WEIGHT config:
# Increase recency bias (recent memories rank higher)
vipune search "authentication" --recency 0.7
# Pure semantic similarity (no recency bias)
vipune search "authentication" --recency 0.0The final score combines semantic similarity and recency time decay:
score = (1 - recency_weight) * similarity + recency_weight * time_score- Default balance: 70% semantic, 30% recency
Apache-2.0 © Janni Turunen