Skip to content

Module 5: AI/ML & Vector Databases


Overview

This module covers AI/ML infrastructure for data platforms, including vector databases for similarity search, LLM Operations (RAG architecture, prompt engineering, embeddings at scale), and feature stores for centralized feature management.


Module Contents

Vector Databases

DocumentDescriptionStatus
Vector Databases OverviewComparison and selection guide✅ Complete
Pinecone GuideManaged vector database✅ Complete
Milvus GuideOpen-source vector database✅ Complete
pgvector GuidePostgreSQL vector extension✅ Complete
Weaviate GuideKnowledge graph vector database✅ Complete

LLM Ops

DocumentDescriptionStatus
LLM Ops OverviewLLM operations overview✅ Complete
RAG ArchitectureRetrieval Augmented Generation✅ Complete
Prompt Engineering PipelinesPrompt management and deployment✅ Complete
Vector Embeddings at ScaleScalable embedding generation✅ Complete

Feature Stores

DocumentDescriptionStatus
Feature Stores OverviewFeature store comparison✅ Complete
Feast GuideOpen-source feature store✅ Complete
Hopsworks GuideEnterprise feature store platform✅ Complete

RAG Architecture Overview

RAG Patterns:

  • Naive RAG: Simple retrieve and generate pipeline
  • Advanced RAG: Query rewriting, reranking, context compression
  • Hybrid Search: Vector + BM25 for best results
  • Metadata Filtering: Pre-filter for better performance

Vector Database Comparison

Feature Comparison

DatabaseTypeDeploymentIndex TypesMax DimensionsCostBest For
PineconeManagedCloud-onlyHNSW20,000$$RAG, production, managed
MilvusOpen SourceSelf-hosted/CloudIVF, HNSW, FLAT32,768$On-premises, K8s
pgvectorExtensionSelf-hostedIVFFlat, HNSW2,000$PostgreSQL shops
WeaviateOpen SourceSelf-hosted/CloudHNSWUnlimited$-$$Knowledge graphs

Selection Criteria


Feature Store Architecture

Feature Store Benefits:

  • Consistency: Same features in training and inference
  • Reusability: Share features across models
  • Version Control: Track feature changes over time
  • Point-in-Time Correctness: Avoid data leakage
  • Governance: Feature ownership, documentation, lineage

Cost Considerations

Vector Database Costs

FactorImpactOptimization
Vector dimensionStorage + computeUse lower dimensions (384 vs. 1536)
Index typeMemory vs. accuracyHNSW for balance
DeploymentOpEx vs. CapExOpen source for scale
RegionData transferColocate with data

Embedding Generation Costs

ModelCost per 1M tokensQualitySpeed
OpenAI ada-002$0.10ExcellentMedium
OpenAI text-embedding-3-small$0.02Very goodFast
Cohere embed-v3$0.10Very goodFast
Sentence TransformersCompute costGoodMedium

Feature Store Costs

PlatformCostNotes
FeastFree (self-hosted)Open source
HopsworksUsage-basedManaged service
TectonCustomEnterprise
Vertex AIUsage-basedGCP integration

Learning Objectives

After completing this module, you will:

  1. Select vector databases: Pinecone vs. Milvus vs. pgvector vs. Weaviate
  2. Implement RAG: Retrieval Augmented Generation patterns
  3. Scale embeddings: Batch and real-time embedding pipelines
  4. Use feature stores: Feast, Hopsworks for ML feature management
  5. Optimize AI/ML costs: Model selection, deployment, serving

Module Dependencies


Quick Start

Vector Databases

  1. Start with vector databases overview for comparison
  2. Choose based on use case:
    • Production RAG: Pinecone
    • Self-hosted: Milvus
    • PostgreSQL shops: pgvector
    • Knowledge graphs: Weaviate

LLM Ops

  1. Learn RAG architecture for LLM systems
  2. Implement prompt engineering pipelines for production
  3. Scale embeddings with batch processing and caching

Feature Stores

  1. Understand feature store patterns
  2. Choose platform:
    • Open-source: Feast
    • Enterprise: Hopsworks

Next Steps

  1. Study Vector Databases
  2. Learn RAG Architecture
  3. Implement Feature Stores
  4. Proceed to Module 6: CI/CD for Data

Estimated Time to Complete Module 5: 8-10 hours

Total Files: 12 markdown files with 60+ diagrams