NVIDIA Blueprintβaligned multi-agent assistant for warehouse operations.
- Overview
- Acronyms & Abbreviations
- System Architecture
- Key Features
- Quick Start
- Multi-Agent System
- API Reference
- Monitoring & Observability
- NeMo Guardrails
- Development Guide
- Contributing
- License
| Acronym | Definition |
|---|---|
| ADR | Architecture Decision Record |
| API | Application Programming Interface |
| BOL | Bill of Lading |
| cuML | CUDA Machine Learning |
| cuVS | CUDA Vector Search |
| EAO | Equipment & Asset Operations (Agent) |
| ERP | Enterprise Resource Planning |
| GPU | Graphics Processing Unit |
| HTTP/HTTPS | Hypertext Transfer Protocol (Secure) |
| IoT | Internet of Things |
| JSON | JavaScript Object Notation |
| JWT | JSON Web Token |
| KPI | Key Performance Indicator |
| LLM | Large Language Model |
| LOTO | Lockout/Tagout |
| MAPE | Mean Absolute Percentage Error |
| MCP | Model Context Protocol |
| NeMo | NVIDIA NeMo |
| NIM/NIMs | NVIDIA Inference Microservices |
| OCR | Optical Character Recognition |
| PPE | Personal Protective Equipment |
| QPS | Queries Per Second |
| RAG | Retrieval-Augmented Generation |
| RAPIDS | Rapid Analytics Platform for Interactive Data Science |
| RBAC | Role-Based Access Control |
| RFID | Radio Frequency Identification |
| RMSE | Root Mean Square Error |
| REST | Representational State Transfer |
| SDS | Safety Data Sheet |
| SKU | Stock Keeping Unit |
| SLA | Service Level Agreement |
| SOP | Standard Operating Procedure |
| SQL | Structured Query Language |
| UI | User Interface |
| UX | User Experience |
| WMS | Warehouse Management System |
This repository implements a production-grade Multi-Agent-Intelligent-Warehouse patterned on NVIDIA's AI Blueprints, featuring:
- Multi-Agent AI System - LangGraph-orchestrated Planner/Router + 5 Specialized Agents (Equipment, Operations, Safety, Forecasting, Document)
- NVIDIA NeMo Integration - Complete document processing pipeline with OCR, structured data extraction, and vision models
- MCP Framework - Model Context Protocol with dynamic tool discovery, execution, and adapter system
- Hybrid RAG Stack - PostgreSQL/TimescaleDB + Milvus vector database with intelligent query routing (90%+ accuracy)
- Production-Grade Vector Search - NV-EmbedQA-E5-v5 embeddings (1024-dim) with NVIDIA cuVS GPU acceleration (19x performance)
- AI-Powered Demand Forecasting - Multi-model ensemble (XGBoost, Random Forest, Gradient Boosting, Ridge, SVR) with NVIDIA RAPIDS GPU acceleration
- Real-Time Monitoring - Equipment status, telemetry, Prometheus metrics, Grafana dashboards, and system health
- Enterprise Security - JWT authentication + RBAC with 5 user roles, NeMo Guardrails for content safety, and comprehensive user management
- System Integrations - WMS (SAP EWM, Manhattan, Oracle), ERP (SAP ECC, Oracle), IoT sensors, RFID/Barcode scanners, Time Attendance systems
- Advanced Features - Redis caching, conversation memory, evidence scoring, intelligent query classification, automated reorder recommendations, business intelligence dashboards
The architecture consists of:
- User/External Interaction Layer - Entry point for users and external systems
- Warehouse Operational Assistant - Central orchestrator managing specialized AI agents
- Agent Orchestration Framework - LangGraph for workflow orchestration + MCP (Model Context Protocol) for tool discovery
- Multi-Agent System - Five specialized agents:
- Equipment & Asset Operations Agent - Equipment assets, assignments, maintenance, and telemetry
- Operations Coordination Agent - Task planning and workflow management
- Safety & Compliance Agent - Safety monitoring, incident response, and compliance tracking
- Forecasting Agent - Demand forecasting, reorder recommendations, and model performance monitoring
- Document Processing Agent - OCR, structured data extraction, and document management
- API Services Layer - Standardized interfaces for business logic and data access
- Data Retrieval & Processing - SQL, Vector, and Knowledge Graph retrievers
- LLM Integration & Orchestration - NVIDIA NIMs with LangGraph orchestration
- Data Storage Layer - PostgreSQL, Vector DB, Knowledge Graph, and Telemetry databases
- Infrastructure Layer - Kubernetes, NVIDIA GPU infrastructure, Edge devices, and Cloud
- Multi-Agent Coordination: LangGraph orchestrates complex workflows between specialized agents
- MCP Integration: Model Context Protocol enables seamless tool discovery and execution
- Hybrid Data Processing: Combines structured (PostgreSQL/TimescaleDB) and vector (Milvus) data
- NVIDIA NIMs Integration: LLM inference and embedding services for intelligent processing
- Real-time Monitoring: Comprehensive telemetry and equipment status tracking
- Scalable Infrastructure: Kubernetes orchestration with GPU acceleration
- Planner/Router - Intelligent query routing and workflow orchestration
- Equipment & Asset Operations Agent - Equipment management, maintenance, and telemetry
- Operations Coordination Agent - Task planning and workflow management
- Safety & Compliance Agent - Safety monitoring and incident response
- Forecasting Agent - Demand forecasting, reorder recommendations, and model performance monitoring
- Document Processing Agent - OCR, structured data extraction, and document management
- MCP Integration - Model Context Protocol with dynamic tool discovery
- Multi-Format Support - PDF, PNG, JPG, JPEG, TIFF, BMP files
- 5-Stage NVIDIA NeMo Pipeline - Complete OCR and structured data extraction
- Real-Time Processing - Background processing with status tracking
- Intelligent OCR -
meta/llama-3.2-11b-vision-instructfor text extraction - Structured Data Extraction - Entity recognition and quality validation
- Hybrid RAG Stack - PostgreSQL/TimescaleDB + Milvus vector database
- Production-Grade Vector Search - NV-EmbedQA-E5-v5 embeddings (1024-dim)
- GPU-Accelerated Search - NVIDIA cuVS-powered vector search (19x performance)
- Intelligent Query Routing - Automatic SQL vs Vector vs Hybrid classification (90%+ accuracy)
- Evidence Scoring - Multi-factor confidence assessment with clarifying questions
- Redis Caching - Intelligent caching with 85%+ hit rate
- π GPU-Accelerated Forecasting - NVIDIA RAPIDS cuML integration for enterprise-scale performance
- 10-100x faster training and inference compared to CPU-only
- Automatic GPU detection - Falls back to CPU if GPU not available
- Full GPU acceleration for Random Forest, Linear Regression, SVR via cuML
- XGBoost GPU support via CUDA when RAPIDS is available
- Seamless integration - No code changes needed, works out of the box
- AI-Powered Demand Forecasting - Multi-model ensemble with Random Forest, XGBoost, Gradient Boosting, Linear Regression, Ridge Regression, SVR
- Advanced Feature Engineering - Lag features, rolling statistics, seasonal patterns, promotional impacts
- Hyperparameter Optimization - Optuna-based tuning with Time Series Cross-Validation
- Real-Time Predictions - Live demand forecasts with confidence intervals
- Automated Reorder Recommendations - AI-suggested stock orders with urgency levels
- Business Intelligence Dashboard - Comprehensive analytics and performance monitoring
- WMS Integration - SAP EWM, Manhattan, Oracle WMS
- ERP Integration - SAP ECC, Oracle ERP
- IoT Integration - Equipment monitoring, environmental sensors, safety systems
- RFID/Barcode Scanning - Honeywell, Zebra, generic scanners
- Time Attendance - Biometric systems, card readers, mobile apps
- Authentication - JWT authentication + RBAC with 5 user roles
- Real-Time Monitoring - Prometheus metrics + Grafana dashboards
- Equipment Telemetry - Battery, temperature, charging analytics
- System Health - Comprehensive observability and alerting
- NeMo Guardrails - Content safety and compliance protection (see NeMo Guardrails section below)
JWT Secret Key Configuration:
- Development: If
JWT_SECRET_KEYis not set, the application uses a default development key with warnings. This allows for easy local development. - Production: The application requires
JWT_SECRET_KEYto be set. If not set or using the default placeholder, the application will fail to start. SetENVIRONMENT=productionand provide a strong, uniqueJWT_SECRET_KEYin your.envfile. - Best Practice: Always set
JWT_SECRET_KEYexplicitly, even in development, using a strong random string (minimum 32 characters).
For more security information, see docs/secrets.md and SECURITY_REVIEW.md.
For complete deployment instructions, see DEPLOYMENT.md.
Option 1: Interactive Jupyter Notebook Setup (Recommended for First-Time Users)
π Complete Setup Guide (Jupyter Notebook)
The interactive notebook provides:
- β Automated environment validation and checks
- β Step-by-step guided setup with explanations
- β Interactive API key configuration
- β Database setup and migration automation
- β User creation and demo data generation
- β Backend and frontend startup from within the notebook
- β Comprehensive error handling and troubleshooting
To use the notebook:
- Open
notebooks/setup/complete_setup_guide.ipynbin Jupyter Lab/Notebook - Follow the interactive cells step by step
- The notebook will guide you through the entire setup process
Option 2: Command-Line Setup (For Experienced Users)
See the Local Development Setup section below for manual command-line setup.
- Python 3.9+ (check with
python3 --version) - Node.js 20.0.0+ (LTS recommended) and npm (check with
node --versionandnpm --version)- Minimum: Node.js 18.17.0+ (required for
node:pathprotocol support) - Recommended: Node.js 20.x LTS for best compatibility
- Note: Node.js 18.0.0 - 18.16.x will fail with
Cannot find module 'node:path'error
- Minimum: Node.js 18.17.0+ (required for
- Docker and Docker Compose
- Git (to clone the repository)
- PostgreSQL client (
psql) - Required for running database migrations- Ubuntu/Debian:
sudo apt-get install postgresql-client - macOS:
brew install postgresqlorbrew install libpq - Windows: Install from PostgreSQL downloads
- Alternative: Use Docker (see DEPLOYMENT.md)
- Ubuntu/Debian:
- Poppler utilities (
poppler-utils) - Required for PDF document processing- Ubuntu/Debian:
sudo apt-get install poppler-utils - macOS:
brew install poppler - Windows: Install from Poppler for Windows or use Chocolatey:
choco install poppler - Note: Required by
pdf2imagepackage for converting PDF pages to images
- Ubuntu/Debian:
- CUDA (for GPU acceleration) - Optional but recommended for RAPIDS GPU-accelerated forecasting
- Recommended: CUDA 12.x (default for RAPIDS packages)
- Supported: CUDA 11.x (via
install_rapids.shauto-detection) - Note: CUDA version is auto-detected during RAPIDS installation. If you have CUDA 13.x, it will install CUDA 12.x packages (backward compatible). For best results, ensure your CUDA driver version matches or exceeds the toolkit version.
For the fastest local development setup:
# 1. Clone repository
git clone https://github.com/NVIDIA-AI-Blueprints/Multi-Agent-Intelligent-Warehouse.git
cd Multi-Agent-Intelligent-Warehouse
# 2. Verify Node.js version (recommended before setup)
./scripts/setup/check_node_version.sh
# 3. Setup environment
./scripts/setup/setup_environment.sh
# 4. Configure environment variables (REQUIRED before starting services)
# Create .env file for Docker Compose (recommended location)
cp .env.example deploy/compose/.env
# Or create in project root: cp .env.example .env
# Edit with your values: nano deploy/compose/.env
# 5. Start infrastructure services
./scripts/setup/dev_up.sh
# 6. Run database migrations
source env/bin/activate
# Load environment variables from .env file (REQUIRED before running migrations)
# This ensures $POSTGRES_PASSWORD is available for the psql commands below
# If .env is in deploy/compose/ (recommended):
set -a && source deploy/compose/.env && set +a
# OR if .env is in project root:
# set -a && source .env && set +a
# Docker Compose: Using Docker Compose (Recommended - no psql client needed)
docker compose -f deploy/compose/docker-compose.dev.yaml exec -T timescaledb psql -U warehouse -d warehouse < data/postgres/000_schema.sql
docker compose -f deploy/compose/docker-compose.dev.yaml exec -T timescaledb psql -U warehouse -d warehouse < data/postgres/001_equipment_schema.sql
docker compose -f deploy/compose/docker-compose.dev.yaml exec -T timescaledb psql -U warehouse -d warehouse < data/postgres/002_document_schema.sql
docker compose -f deploy/compose/docker-compose.dev.yaml exec -T timescaledb psql -U warehouse -d warehouse < data/postgres/004_inventory_movements_schema.sql
docker compose -f deploy/compose/docker-compose.dev.yaml exec -T timescaledb psql -U warehouse -d warehouse < scripts/setup/create_model_tracking_tables.sql
# 7. Create default users
python scripts/setup/create_default_users.py
# 8. Generate demo data (optional but recommended)
python scripts/data/quick_demo_data.py
# 9. Generate historical demand data for forecasting (optional, required for Forecasting page)
python scripts/data/generate_historical_demand.py
# 10. (Optional) Install RAPIDS GPU acceleration for forecasting
# This enables 10-100x faster forecasting with NVIDIA GPUs
# Requires: NVIDIA GPU with CUDA 12.x support
./scripts/setup/install_rapids.sh
# Or manually: pip install --extra-index-url=https://pypi.nvidia.com cudf-cu12 cuml-cu12
# 11. Start API server
./scripts/start_server.sh
# 12. Start frontend (in another terminal)
cd src/ui/web
npm install
npm startAccess:
- Frontend: http://localhost:3001 (login:
admin/changeme) - API: http://localhost:8001
- API Docs: http://localhost:8001/docs
Service Endpoints:
- Postgres/Timescale:
postgresql://warehouse:changeme@localhost:5435/warehouse - Redis:
localhost:6379 - Milvus gRPC:
localhost:19530 - Kafka:
localhost:9092
.env file location matters!
Docker Compose looks for .env files in this order:
- Same directory as the compose file (
deploy/compose/.env) - Current working directory (project root
.env)
Recommended: Create .env in the same directory as your compose file for consistency:
# Option 1: In deploy/compose/ (recommended for Docker Compose)
cp .env.example deploy/compose/.env
nano deploy/compose/.env # or your preferred editor
# Option 2: In project root (works if running commands from project root)
cp .env.example .env
nano .env # or your preferred editorCritical Variables:
- Database connection settings (POSTGRES_USER, POSTGRES_PASSWORD, POSTGRES_DB, DB_HOST, DB_PORT)
- Redis connection (REDIS_HOST, REDIS_PORT)
- Milvus connection (MILVUS_HOST, MILVUS_PORT)
- JWT secret key (JWT_SECRET_KEY) - Required in production. In development, a default is used with warnings. See Security Notes below.
- Admin password (DEFAULT_ADMIN_PASSWORD)
For AI Features (Optional):
- NVIDIA API keys (NVIDIA_API_KEY, NEMO_API_KEY, LLAMA_API_KEY)
Quick Setup for NVIDIA API Keys:
python setup_nvidia_api.pyNode.js Version Issues:
- Error: "Cannot find module 'node:path'": Your Node.js version is too old
- Check version:
node --version - Minimum required: Node.js 18.17.0+
- Recommended: Node.js 20.x LTS
- Run version check:
./scripts/setup/check_node_version.sh - Upgrade:
nvm install 20 && nvm use 20(if using nvm) - Or download from: https://nodejs.org/
- After upgrading, clear and reinstall:
cd src/ui/web && rm -rf node_modules package-lock.json && npm install
- Check version:
Database Connection Issues:
- Ensure Docker containers are running:
docker ps - Check TimescaleDB logs:
docker logs wosa-timescaledb - Verify port 5435 is not in use
API Server Won't Start:
- Ensure virtual environment is activated:
source env/bin/activate - Check Python version:
python3 --version(must be 3.9+) - Use the startup script:
./scripts/start_server.sh - See DEPLOYMENT.md troubleshooting section
Frontend Build Issues:
- Verify Node.js version:
./scripts/setup/check_node_version.sh - Clear node_modules:
cd src/ui/web && rm -rf node_modules package-lock.json && npm install - Check for port conflicts: Ensure port 3001 is available
For more help: See DEPLOYMENT.md for detailed troubleshooting or open an issue on GitHub.
The Warehouse Operational Assistant uses a sophisticated multi-agent architecture with specialized AI agents for different aspects of warehouse operations.
Mission: Ensure equipment is available, safe, and optimally used for warehouse workflows.
Key Capabilities:
- Equipment assignment and tracking
- Real-time telemetry monitoring (battery, temperature, charging status)
- Maintenance management and scheduling
- Asset tracking and location monitoring
- Equipment utilization analytics
Action Tools: assign_equipment, get_equipment_status, create_maintenance_request, get_equipment_telemetry, update_equipment_location, get_equipment_utilization, create_equipment_reservation, get_equipment_history
Mission: Coordinate warehouse operations, task planning, and workflow optimization.
Key Capabilities:
- Task management and assignment
- Workflow optimization (pick paths, resource allocation)
- Performance monitoring and KPIs
- Resource planning and allocation
Action Tools: create_task, assign_task, optimize_pick_path, get_task_status, update_task_progress, get_performance_metrics, create_work_order, get_task_history
Mission: Ensure warehouse safety compliance and incident management.
Key Capabilities:
- Incident management and logging
- Safety procedures and checklists
- Compliance monitoring and training
- Emergency response coordination
Action Tools: log_incident, start_checklist, broadcast_alert, create_corrective_action, lockout_tagout_request, near_miss_capture, retrieve_sds
Mission: Provide AI-powered demand forecasting, reorder recommendations, and model performance monitoring.
Key Capabilities:
- Demand forecasting using multiple ML models
- Automated reorder recommendations with urgency levels
- Model performance monitoring (accuracy, MAPE, drift scores)
- Business intelligence and trend analysis
- Real-time predictions with confidence intervals
Action Tools: get_forecast, get_batch_forecast, get_reorder_recommendations, get_model_performance, get_forecast_dashboard, get_business_intelligence
Forecasting Models:
- Random Forest (82% accuracy, 15.8% MAPE)
- XGBoost (79.5% accuracy, 15.0% MAPE)
- Gradient Boosting (78% accuracy, 14.2% MAPE)
- Linear Regression, Ridge Regression, SVR
Model Availability by Phase:
| Model | Phase 1 & 2 | Phase 3 |
|---|---|---|
| Random Forest | β | β |
| XGBoost | β | β |
| Time Series | β | β |
| Gradient Boosting | β | β |
| Ridge Regression | β | β |
| SVR | β | β |
| Linear Regression | β | β |
Mission: Process warehouse documents with OCR and structured data extraction.
Key Capabilities:
- Multi-format document support (PDF, PNG, JPG, JPEG, TIFF, BMP)
- Intelligent OCR with NVIDIA NeMo
- Structured data extraction (invoices, receipts, BOLs)
- Quality assessment and validation
All agents are integrated with the Model Context Protocol (MCP) framework:
- Dynamic Tool Discovery - Real-time tool registration and discovery
- Cross-Agent Communication - Seamless tool sharing between agents
- Intelligent Routing - MCP-enhanced intent classification
- Tool Execution Planning - Context-aware tool execution
See docs/architecture/mcp-integration.md for detailed MCP documentation.
GET /api/v1/health- System health checkGET /api/v1/health/simple- Simple health statusGET /api/v1/version- API version information
POST /api/v1/auth/login- User authenticationGET /api/v1/auth/me- Get current user informationGET /api/v1/auth/users/public- Get list of users for dropdown selection (public, no auth required)GET /api/v1/auth/users- Get all users (admin only)
POST /api/v1/chat- Chat with multi-agent system (requires NVIDIA API keys)
GET /api/v1/equipment- List all equipmentGET /api/v1/equipment/{asset_id}- Get equipment detailsGET /api/v1/equipment/{asset_id}/status- Get equipment statusGET /api/v1/equipment/{asset_id}/telemetry- Get equipment telemetryGET /api/v1/equipment/assignments- Get equipment assignmentsGET /api/v1/equipment/maintenance/schedule- Get maintenance schedulePOST /api/v1/equipment/assign- Assign equipmentPOST /api/v1/equipment/release- Release equipmentPOST /api/v1/equipment/maintenance- Schedule maintenance
GET /api/v1/forecasting/dashboard- Comprehensive forecasting dashboardGET /api/v1/forecasting/real-time- Real-time demand predictionsGET /api/v1/forecasting/reorder-recommendations- Automated reorder suggestionsGET /api/v1/forecasting/model-performance- Model performance metricsGET /api/v1/forecasting/business-intelligence- Business analyticsPOST /api/v1/forecasting/batch-forecast- Batch forecast for multiple SKUsGET /api/v1/training/history- Training historyPOST /api/v1/training/start- Start model training
POST /api/v1/document/upload- Upload document for processingGET /api/v1/document/status/{document_id}- Check processing statusGET /api/v1/document/results/{document_id}- Get extraction resultsGET /api/v1/document/analytics- Document analytics
GET /api/v1/operations/tasks- List tasksGET /api/v1/safety/incidents- List safety incidents
Full API Documentation: http://localhost:8001/docs (Swagger UI)
The system includes comprehensive monitoring with Prometheus metrics collection and Grafana dashboards.
Quick Start:
# Start monitoring stack
./scripts/setup/setup_monitoring.shSee DEPLOYMENT.md for detailed monitoring setup instructions.
Access URLs:
- Grafana: http://localhost:3000 (admin/changeme)
- Prometheus: http://localhost:9090
- Alertmanager: http://localhost:9093
Key Metrics Tracked:
- API request rates and latencies
- Equipment telemetry and status
- Agent performance and response times
- Database query performance
- Vector search performance
- Cache hit rates and memory usage
See monitoring/ for dashboard configurations and alerting rules.
The system implements NVIDIA NeMo Guardrails for content safety, security, and compliance protection. All user inputs and AI responses are validated through a comprehensive guardrails system to ensure safe and compliant interactions.
The guardrails system provides dual implementation support with automatic fallback:
- NeMo Guardrails SDK (with Colang) - Intelligent, programmable guardrails using NVIDIA's official SDK
- β
Already included in
requirements.txt(nemoguardrails>=0.19.0) - Installed automatically when you run
pip install -r requirements.txt
- β
Already included in
- Pattern-Based Matching - Fast, lightweight fallback using keyword/phrase matching
- Feature Flag Control - Runtime switching between implementations via
USE_NEMO_GUARDRAILS_SDK - Automatic Fallback - Seamlessly switches to pattern-based if SDK unavailable
- Input & Output Validation - Checks both user queries and AI responses
- Timeout Protection - Prevents hanging requests (3s input, 5s output)
- Comprehensive Monitoring - Metrics tracking for method usage and performance
The guardrails system protects against 88 patterns across 5 categories:
- Jailbreak Attempts (17 patterns) - Prevents instruction override attempts
- Safety Violations (13 patterns) - Blocks unsafe operational guidance
- Security Violations (15 patterns) - Prevents security information requests
- Compliance Violations (12 patterns) - Ensures regulatory adherence
- Off-Topic Queries (13 patterns) - Redirects non-warehouse queries
# Enable SDK implementation (recommended)
USE_NEMO_GUARDRAILS_SDK=true
# NVIDIA API key (required for SDK)
NVIDIA_API_KEY=your-api-key-here
# Optional: Guardrails-specific configuration
RAIL_API_KEY=your-api-key-here # Falls back to NVIDIA_API_KEY if not set
RAIL_API_URL=https://integrate.api.nvidia.com/v1
GUARDRAILS_TIMEOUT=10
GUARDRAILS_USE_API=trueGuardrails are automatically integrated into the chat endpoint:
- Input Safety Check - Validates user queries before processing (3s timeout)
- Output Safety Check - Validates AI responses before returning (5s timeout)
- Metrics Tracking - Logs method used, performance, and safety status
# Unit tests
pytest tests/unit/test_guardrails_sdk.py -v
# Integration tests (compares both implementations)
pytest tests/integration/test_guardrails_comparison.py -v -s
# Performance benchmarks
pytest tests/integration/test_guardrails_comparison.py::test_performance_benchmark -v -sπ For comprehensive documentation, see: Guardrails Implementation Guide
The detailed guide includes:
- Complete architecture overview
- Implementation details (SDK vs Pattern-based)
- All 88 guardrails patterns
- API interface documentation
- Configuration reference
- Monitoring & metrics
- Testing instructions
- Troubleshooting guide
- Future roadmap
Key Files:
- Service:
src/api/services/guardrails/guardrails_service.py - SDK Wrapper:
src/api/services/guardrails/nemo_sdk_service.py - Colang Config:
data/config/guardrails/rails.co - NeMo Config:
data/config/guardrails/config.yml - Legacy YAML:
data/config/guardrails/rails.yaml
.
ββ src/ # Source code
β ββ api/ # FastAPI application
β ββ retrieval/ # Retrieval services
β ββ memory/ # Memory services
β ββ adapters/ # External system adapters
β ββ ui/ # React web dashboard
ββ data/ # SQL DDL/migrations, sample data
ββ deploy/ # Deployment configurations
β ββ compose/ # Docker Compose files
β ββ helm/ # Helm charts
β ββ scripts/ # Deployment scripts
ββ scripts/ # Utility scripts
β ββ setup/ # Setup scripts
β ββ forecasting/ # Forecasting scripts
β ββ data/ # Data generation scripts
ββ tests/ # Test suite
ββ docs/ # Documentation
β ββ architecture/ # Architecture documentation
ββ monitoring/ # Prometheus/Grafana configs
API Server:
source env/bin/activate
./scripts/start_server.shFrontend:
cd src/ui/web
npm startInfrastructure:
./scripts/setup/dev_up.sh# Run all tests
pytest tests/
# Run specific test suite
pytest tests/unit/
pytest tests/integration/- Architecture: docs/architecture/
- MCP Integration: docs/architecture/mcp-integration.md
- Forecasting: docs/forecasting/
- Deployment: DEPLOYMENT.md - Complete deployment guide with Docker and Kubernetes options
Contributions are welcome! Please see our contributing guidelines and code of conduct.
- Fork the repository
- Create a feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'feat: add amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
Commit Message Format: We use Conventional Commits:
feat:- New featurefix:- Bug fixdocs:- Documentation changesrefactor:- Code refactoringtest:- Test additions/changes
See LICENSE for license information.
Governing Terms: The Blueprint scripts are governed by Apache License, Version 2.0, and enables use of separate open source and proprietary software governed by their respective licenses: Llama-3.3-nemotron-super-49b-v1.5, NVIDIA Retrieval QA Llama 3.2 1B Embedding v2, NeMo Retriever Extraction, NeMo Retriever Page Elements v3, NeMo Retriever OCR v1, Nemotron Parse, NVIDIA-Nemotron-Nano-12B-v2-VL, NeMo Guardrails, and RAPIDS cuML.
This project will download and install additional 3rd party open source software projects. Review the license terms for these open source projects before use.
