W

rag-implementation

by wshobson

Build Retrieval-Augmented Generation (RAG) systems for LLM applications with vector databases and semantic search. Use when implementing knowledge-grounded AI, building document Q&A systems, or integrating LLMs with external knowledge bases.

Stars0
Favorites0
Comments0
AddedMar 28, 2026
CategoryRAG Workflows
Install Command
npx skills add https://github.com/wshobson/agents --skill rag-implementation
Overview

Overview

What is rag-implementation?

rag-implementation is a practical skill for building Retrieval-Augmented Generation (RAG) workflows in large language model (LLM) applications. It enables you to connect LLMs with external knowledge sources using vector databases and semantic search, resulting in more accurate, grounded, and up-to-date responses.

Who should use this skill?

This skill is ideal for developers, data scientists, and AI engineers who need to:

  • Build Q&A systems over proprietary or internal documents
  • Create chatbots that deliver current, factual information
  • Implement semantic search with natural language queries
  • Reduce hallucinations in LLM outputs by grounding responses in real data
  • Enable LLMs to access domain-specific or private knowledge bases
  • Develop documentation assistants or research tools with source citation

Problems solved

rag-implementation addresses the challenge of connecting LLMs to external knowledge, allowing for:

  • Accurate, context-aware answers
  • Retrieval of relevant documents or passages
  • Integration with modern vector databases and embedding models

How to Use

Installation Steps

  1. Install the skill using the following command:
    npx skills add https://github.com/wshobson/agents --skill rag-implementation
    
  2. Begin by reviewing the SKILL.md file for a high-level overview and usage guidance.
  3. Explore supporting files such as README.md, AGENTS.md, and metadata.json for deeper context. Check any rules/, resources/, references/, or scripts/ directories if present.

Core Components

Vector Databases

rag-implementation supports integration with leading vector databases for efficient storage and retrieval of document embeddings. Popular options include:

  • Pinecone (managed, scalable)
  • Weaviate (open-source, hybrid search)
  • Milvus (high performance, on-premise)
  • Chroma (lightweight, local development)
  • Qdrant (fast, Rust-based)
  • pgvector (PostgreSQL extension)

Embedding Models

Convert text into numerical vectors for semantic search using models such as:

  • voyage-3-large (Anthropic/Claude apps)
  • voyage-code-3 (code search)
  • text-embedding-3-large (OpenAI apps)

Adapting the Workflow

Rather than copying the workflow verbatim, adapt the provided structure to your own repository, tools, and operational requirements. This ensures compatibility with your data sources and LLM stack.

FAQ

What does rag-implementation provide out of the box?

It offers a structured approach to building RAG pipelines, including guidance on selecting vector databases and embedding models, and best practices for integrating LLMs with external knowledge.

When should I use rag-implementation?

Use this skill when you need to ground LLM outputs in proprietary, current, or domain-specific data, such as for document Q&A, semantic search, or research tools.

What files should I review first?

Start with SKILL.md for an overview. Then check README.md and any supporting files for implementation details.

Where can I find more details?

Open the Files tab in the repository to explore the full file tree, including references and helper scripts for advanced customization.

Ratings & Reviews

No ratings yet
Share your review
Sign in to leave a rating and comment for this skill.
G
0/10000
Latest reviews
Saving...