Marketplace rag-pipeline
Details on the Retrieval Augmented Generation pipeline, Ingestion, and Vector Search.
install
source · Clone the upstream repo
git clone https://github.com/aiskillstore/marketplace
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/aiskillstore/marketplace "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/abdulsamad94/rag-pipeline" ~/.claude/skills/aiskillstore-marketplace-rag-pipeline && rm -rf "$T"
manifest:
skills/abdulsamad94/rag-pipeline/SKILL.mdsource content
RAG Pipeline Logic
Ingestion
- Script:
backend/ingest.py - Process:
- Scans
.docs/ - Cleans MDX (removes frontmatter/imports).
- Chunks text (1000 chars, 100 overlap).
- Embeds using
.models/text-embedding-004 - Upserts to Qdrant collection
.physical_ai_book
- Scans
- Run:
python backend/ingest.py
Vector Search (Qdrant)
- Client:
qdrant-client - Collection:
physical_ai_book - Vector Size: 768 (Gecko-004)
- Similarity: Cosine
Prompt Engineering
- File:
.backend/utils/helpers.py - RAG Prompt: Constructs a prompt containing retrieved context chunks.
- Personalization:
creates system instructions based onbackend/personalization.py
andsoftware_background
of the user.hardware_background
Agentic Flow
We use a custom
Agent class (backend/agents.py) that wraps the LLM calls, allowing for future expansion into multi-agent workflows.