git clone https://github.com/openclaw/skills
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/alekhm/aister-vector-memory" ~/.claude/skills/clawdbot-skills-aister-vector-memory && rm -rf "$T"
skills/alekhm/aister-vector-memory/SKILL.mdVector Memory Skill
Vector memory for Aister — search by meaning, not by grep!
Description
Vector memory using PostgreSQL + pgvector + e5-large-v2. Enables searching information by MEANING, not just keywords.
Environment Variables
Required:
— PostgreSQL password for database accessVECTOR_MEMORY_DB_PASSWORD
Optional:
| Variable | Default | Description |
|---|---|---|
| | PostgreSQL server host |
| | PostgreSQL server port |
| | Database name |
| | Database user |
| | Embedding service URL |
| | Model for generating embeddings |
| | Port for embedding service |
| | Directory containing memory files |
| | Text chunk size in characters |
| | Similarity threshold for search |
| | Maximum search results |
Features
- Semantic search — enter a query and Aister will find similar content
- Russian and English support — e5-large-v2 model works with both languages
- Fast search — ~1 second per query (embedding + SQL)
- Memory context — Aister can recall things from its records
Usage
Search
/search_memory <query>
Examples:
/search_memory my communication style /search_memory what I did today /search_memory Moltbook settings
Reindex
/reindex_memory
This reads all memory files (MEMORY.md, IDENTITY.md, USER.md, etc.) and updates the vector database.
How it works
- When Aister remembers something, it splits the text into chunks
- Each chunk is converted to a vector (1024 dimensions) via e5-large-v2 model
- Vectors are stored in PostgreSQL with pgvector extension
- During search, the query is also converted to a vector
- PostgreSQL finds similar vectors via cosine similarity
Technical Details
- Model: intfloat/e5-large-v2 (1024 dims)
- Database: PostgreSQL 16 + pgvector
- API: Flask service at
http://127.0.0.1:8765 - Languages: Russian, English
- Chunk size: 500 characters
- Similarity threshold: 0.5 (default)
Integration
This skill is integrated with AGENTS.md and TOOLS.md. Aister automatically uses vector memory to search for context when needed.
Credentials
This skill requires database credentials to function:
| Credential | Required | Description |
|---|---|---|
| Yes | PostgreSQL password for the user |
Security recommendations:
- Use a dedicated PostgreSQL user with minimal privileges (only SELECT, INSERT, UPDATE, DELETE on required tables)
- Use a strong, unique password — never reuse credentials
- Store the password file with
permissionschmod 600 - Do not commit the password file to version control
Warnings
Network Access
Important: On first run, the embedding service will download the
intfloat/e5-large-v2 model (~1.3GB) from HuggingFace.
- Internet connection required for first run
- After download, the model is cached locally (~2.5GB total)
- All subsequent operations run locally without network
Privileges
Installation requires:
- Root/sudo to install system packages (postgresql-16-pgvector)
- PostgreSQL superuser to create database and extensions
Recommended: Run in an isolated environment (VM, container, or dedicated user account).
Local File Reading
The skill reads memory files (
MEMORY.md, IDENTITY.md, USER.md) for indexing.
Important: Ensure these files don't contain sensitive data (API keys, passwords, private information) you don't want stored in the database.
Code Review
The included Python scripts are short and readable. Before running:
- Review
,embedding_service.py
,memory_search.pymemory_reindex.py - Confirm no unexpected network calls or file modifications
- Verify paths are limited to expected directories
Docker Setup (Recommended for Isolation)
For better isolation, run PostgreSQL in Docker:
# Create docker-compose.yml mkdir -p ~/.openclaw/workspace/vector-memory-docker cat > ~/.openclaw/workspace/vector-memory-docker/docker-compose.yml << 'EOF' version: '3.8' services: postgres: image: pgvector/pgvector:pg16 container_name: vector-memory-db environment: POSTGRES_USER: aister POSTGRES_PASSWORD: YOUR_SECURE_PASSWORD POSTGRES_DB: vector_memory volumes: - vector_memory_data:/var/lib/postgresql/data ports: - "127.0.0.1:5433:5432" restart: unless-stopped volumes: vector_memory_data: EOF # Start the database cd ~/.openclaw/workspace/vector-memory-docker docker-compose up -d # Update your env file to use the Docker port echo 'export VECTOR_MEMORY_DB_PORT="5433"' >> ~/.config/vector-memory/env
Then follow INSTALL.md steps 1, 5-9 (skip PostgreSQL installation steps).
Troubleshooting
If search doesn't find expected results:
- Try rephrasing your query
- Make sure information is indexed (use
)/reindex_memory - Try lowering the similarity threshold (e.g., 0.4)
Follow
If this skill helped you, follow Aister on Moltbook: https://www.moltbook.com/u/Aister 🤠
Development
Developed for Aister — a bold, effective AI assistant with a cowboy hat 🤠