prvng_core/scripts/ai_demo.nu

73 lines
2.6 KiB
Plaintext
Raw Permalink Normal View History

#!/usr/bin/env nu
# AI Integration Demo Script
print "🤖 AI Integration for Infrastructure Automation"
print "==============================================="
print ""
print "✅ AI Implementation Status:"
print " 1. Nickel Configuration Schema: nickel/settings.ncl:54-130"
print " 2. Core AI Library: core/nulib/lib_provisioning/ai/lib.nu"
print " 3. Template Generation: Enhanced with AI prompts"
print " 4. Natural Language Queries: --ai_query flag added"
print " 5. Webhook Integration: Chat platform support"
print " 6. CLI Integration: AI command module implemented"
print ""
print "🔧 Configuration Required:"
print " Set API key environment variable:"
print " - export OPENAI_API_KEY='your-key' (for OpenAI)"
print " - export ANTHROPIC_API_KEY='your-key' (for Claude)"
print " - export LLM_API_KEY='your-key' (for generic LLM)"
print ""
print " Enable in Nickel settings:"
print " ai: AIProvider {"
print " enabled: true"
print " provider: \"openai\" # or \"claude\" or \"generic\""
print " max_tokens: 2048"
print " temperature: 0.3"
print " enable_template_ai: true"
print " enable_query_ai: true"
print " enable_webhook_ai: false"
print " }"
print ""
print "📋 Usage Examples (once configured):"
print ""
print " # Generate infrastructure templates"
print " ./core/nulib/provisioning ai template \\"
print " --prompt \"3-node Kubernetes cluster with Ceph storage\""
print ""
print " # Natural language queries"
print " ./core/nulib/provisioning query \\"
print " --ai_query \"show all AWS servers with high CPU usage\""
print ""
print " # Test AI connectivity"
print " ./core/nulib/provisioning ai test"
print ""
print " # Show AI configuration"
print " ./core/nulib/provisioning ai config"
print ""
print "🌟 Key Features:"
print " - Optional running mode (disabled by default)"
print " - Multiple provider support (OpenAI, Claude, generic LLM)"
print " - Template generation from natural language"
print " - Infrastructure queries in plain English"
print " - Chat platform integration (Slack, Discord, Teams)"
print " - Context-aware responses"
print " - Configurable per feature (template, query, webhook)"
print ""
print "🔒 Security:"
print " - API keys via environment variables only"
print " - No secrets stored in configuration files"
print " - Optional webhook AI (disabled by default)"
print " - Validate all AI-generated configurations"
print ""
print "🎯 Implementation Complete!"
print " All requested AI capabilities have been integrated as optional features"
print " with support for OpenAI, Claude, and generic LLM providers."