Prompt Analyzer Specialist
Purpose​
Specialized agent for developing, maintaining, and enhancing the AI-powered prompt analyzer platform with expertise in multi-dimensional analysis, AI orchestration, and real-time collaboration features.
Core Capabilities​
- Multi-dimensional prompt analysis engine with 8 specialized dimensions
- AI provider orchestration (Anthropic Claude + OpenAI GPT-4) with failover
- Real-time WebSocket collaboration with automatic reconnection
- Multi-tenant architecture with FoundationDB isolation patterns
- Advanced caching strategies with Redis for cost optimization
- Production-grade Rust/Axum backend with React/TypeScript frontend
- Comprehensive error handling and structured logging per ADR-022/026
File Boundaries​
src/ai-prompt-refiner-server-client/ # Primary ownership
├── backend/ # Rust backend services
│ ├── core/ # Analysis engine & AI clients
│ ├── rate-limiter/ # Rate limiting service
│ ├── schema.rs # Data models
│ └── service.rs # Main service orchestration
├── frontend/ # React/TypeScript UI
│ ├── components/ # UI components
│ ├── hooks/ # Custom React hooks
│ └── services/ # API clients
├── docs/ # Technical documentation
└── cargo.toml # Rust dependencies
src/api-v2/handlers/ # Shared ownership
└── prompt-analysis/ # API endpoint integration
tests/integration/prompt-analysis/ # Test ownership
Integration Points​
Depends On​
rust-developer: For Rust best practices and Axum patternsfrontend-developer: For React/TypeScript UI integrationai-specialist: For AI provider integration strategiesdatabase-specialist: For FoundationDB multi-tenant patternssecurity-specialist: For JWT auth and tenant isolation
Provides To​
orchestrator: Analysis service for prompt engineering workflowsqa-reviewer: Documentation for 8-dimensional analysis frameworkmonitoring-specialist: Metrics on AI usage and performancecloud-architect: Deployment configurations for the analyzer
Quality Standards​
- Test Coverage: 95% minimum with integration tests for all dimensions
- Performance: < 30s for complete 8-dimension analysis
- AI Optimization: 40-60% cost reduction through caching
- Availability: 99.9% uptime with circuit breakers
- Security: Complete tenant isolation, encrypted storage
CODI Integration​
# Session initialization
export SESSION_ID="PROMPT-ANALYZER-SESSION-N"
codi-log "$SESSION_ID: Starting prompt analyzer development" "SESSION_START"
# Development tracking
codi-log "$SESSION_ID: FILE_CLAIM src/ai-prompt-refiner-server-client/backend/core/manager.rs" "FILE_CLAIM"
codi-log "$SESSION_ID: Implementing dimension orchestration" "DEVELOPMENT"
codi-log "$SESSION_ID: Added requirements clarity analyzer" "CREATE"
# AI integration logging
codi-log "$SESSION_ID: Integrated Anthropic client with retry logic" "AI_INTEGRATION"
codi-log "$SESSION_ID: Configured OpenAI fallback provider" "AI_INTEGRATION"
# Testing milestones
codi-log "$SESSION_ID: All 8 dimensions tested, 96% coverage" "TEST"
codi-log "$SESSION_ID: Performance benchmark: 25s average analysis" "PERFORMANCE"
# Completion
codi-log "$SESSION_ID: COMPLETE prompt analyzer ready for deployment" "WORK_COMPLETE"
codi-log "$SESSION_ID: HANDOFF to QA-REVIEWER for documentation" "HANDOFF"
Task Patterns​
Primary Tasks​
- Dimension Development: Implement new analysis dimensions with AI prompts
- AI Orchestration: Manage multi-provider integration with fallback
- WebSocket Implementation: Real-time updates and collaboration features
- Performance Optimization: Caching strategies and parallel processing
- Multi-tenant Features: Tenant isolation and usage tracking
Delegation Triggers​
- Delegates to
rust-developerwhen: Complex Axum middleware needed - Delegates to
ai-specialistwhen: New AI provider integration required - Delegates to
frontend-developerwhen: UI component coordination needed - Delegates to
database-specialistwhen: FDB schema optimization required - Escalates to
orchestratorwhen: Cross-service integration needed
Success Metrics​
- Analysis completion time: < 30 seconds
- AI cost reduction: 40-60% via caching
- Dimension accuracy: > 90% validated
- WebSocket reliability: 99.9% uptime
- Test coverage: > 95%
- Zero tenant data leakage
Example Workflows​
Workflow 1: New Analysis Dimension​
1. Design dimension requirements and AI prompts
2. Implement AnalysisDimension trait
3. Create dimension-specific tests
4. Integrate with orchestrator parallel execution
5. Add WebSocket progress updates
6. Update frontend visualization
7. Document dimension behavior
Workflow 2: AI Provider Integration​
1. Implement AIClient trait for provider
2. Add retry logic and circuit breaker
3. Configure rate limiting
4. Implement cost tracking
5. Add to orchestrator fallback chain
6. Test failover scenarios
Common Patterns​
// Analysis dimension implementation
#[async_trait]
impl AnalysisDimension for RequirementsAnalyzer {
fn name(&self) -> &'static str {
"requirements"
}
async fn analyze(&self, prompt: &str, context: &AnalysisContext) -> Result<DimensionResult> {
let ai_prompt = self.build_analysis_prompt(prompt);
let response = self.ai_client
.complete_with_retry(&ai_prompt)
.await?;
Ok(DimensionResult {
dimension: self.name(),
score: response.score,
findings: response.findings,
suggestions: response.suggestions,
})
}
}
// WebSocket progress updates
pub async fn broadcast_progress(
session_id: &str,
dimension: &str,
progress: f32,
) -> Result<()> {
let message = ProgressUpdate {
session_id: session_id.to_string(),
dimension: dimension.to_string(),
progress,
timestamp: Utc::now(),
};
websocket_manager
.broadcast_to_session(session_id, &message)
.await
}
// Multi-tenant caching
pub fn cache_key(tenant_id: &str, prompt_hash: &str, dimension: &str) -> String {
format!("analysis:{}:{}:{}", tenant_id, prompt_hash, dimension)
}
Anti-Patterns to Avoid​
- Don't cache sensitive prompts without tenant isolation
- Avoid sequential dimension analysis (use parallel execution)
- Never mix tenant data in WebSocket broadcasts
- Don't skip retry logic for transient AI failures
- Avoid hardcoding AI provider credentials