Test Knowledge
Learn how to validate and test your knowledge bases to ensure they provide accurate and relevant information for your AI workflows.
Overview
Testing your knowledge base ensures:
- Accurate Retrieval: Content is found when needed
- Relevant Results: Search returns appropriate information
- Quality Validation: Content is processed correctly
- Performance Verification: Search responds quickly
- Workflow Readiness: Knowledge works effectively in automation
Knowledge Testing Interface
Accessing the Test Feature
- Navigate to Knowledge: Go to your knowledge base
- Test Tab: Click on "Test" or "Manual Run" tab
- Test Interface: Access the knowledge retrieval testing interface
Test Interface Components
Search Parameters Section
- Query Input: Enter search questions or keywords
- Max Results: Control number of results returned
- Search Options: Advanced search parameters
- Filter Controls: Content filtering options
Results Display
- Result List: Search results with relevance scores
- Content Preview: Preview of retrieved content
- Source Information: Original document and location
- Metadata Display: Document properties and processing info
Manual Testing Process
Basic Search Testing
Step 1: Simple Queries Start with straightforward questions:
Example Queries:
- "What are the office hours?"
- "How do I reset my password?"
- "What is the return policy?"
- "System requirements"
Step 2: Review Results
- Relevance: Are results related to your query?
- Completeness: Do results contain the information you need?
- Accuracy: Is the information correct and up-to-date?
- Source Quality: Are results from appropriate documents?
Step 3: Complex Queries Test more sophisticated searches:
Complex Query Examples:
- "How do I configure SSL certificates for the web server?"
- "What are the differences between Basic and Premium plans?"
- "Troubleshooting steps for database connection errors"
- "Process for handling customer refund requests"
Query Variations Testing
Different Phrasings Test the same question in multiple ways:
Password Reset Example:
- "How do I reset my password?"
- "Password reset process"
- "Forgot password procedure"
- "Change password steps"
- "Reset login credentials"
Keyword vs. Natural Language
- Keywords: "SSL configuration Apache"
- Natural Language: "How do I set up SSL for Apache web server?"
- Mixed: "Apache SSL setup process"
Specific vs. General
- Specific: "Windows 10 installation requirements"
- General: "system requirements"
- Context: "What do I need to install this software?"
Advanced Testing Scenarios
Edge Case Testing
Misspellings and Typos Test how knowledge handles common errors:
Misspelling Examples:
- "passowrd reset" (password)
- "instalation guide" (installation)
- "configurtion steps" (configuration)
Synonyms and Alternatives Test different terminology:
Synonym Examples:
- "bug" vs "issue" vs "problem"
- "setup" vs "configuration" vs "installation"
- "user" vs "customer" vs "client"
Partial Information Test with incomplete queries:
Partial Query Examples:
- "SSL" (expecting SSL configuration info)
- "refund" (expecting refund policy)
- "backup" (expecting backup procedures)
Content Coverage Testing
Document Verification Ensure all important documents are indexed:
- List Key Documents: Identify critical content
- Test Coverage: Search for information from each document
- Verify Retrieval: Confirm content is found and accurate
- Check Processing: Validate text extraction quality
Topic Coverage Test different subject areas:
- Technical Topics: API documentation, configuration guides
- Business Topics: Policies, procedures, guidelines
- User Topics: How-to guides, troubleshooting, FAQs
- Reference Topics: Specifications, standards, definitions
Performance Testing
Response Time
- Simple Queries: Should return results quickly (under 1 second)
- Complex Queries: May take longer but should be reasonable
- Large Result Sets: Test with queries that return many results
- Concurrent Testing: Test multiple simultaneous queries
Result Quality vs. Speed
- Accuracy: More results may be more accurate but slower
- Relevance: Fewer, more relevant results may be faster
- Optimization: Find the right balance for your use case
Testing Different Content Types
Structured Documents
Technical Documentation
Test Queries:
- "API authentication methods"
- "Database schema requirements"
- "System architecture overview"
Policies and Procedures
Test Queries:
- "Employee onboarding process"
- "Expense reimbursement policy"
- "Security incident response"
Unstructured Content
Text Documents
- Test extraction from various text formats
- Verify handling of formatting and structure
- Check processing of headers and sections
Web Content
- Test crawled web page content
- Verify link and navigation content removal
- Check handling of dynamic content
Mixed Content
- Test documents with text, tables, and images
- Verify handling of complex layouts
- Check extraction from presentations and spreadsheets
Automated Testing
Test Scripts
Create repeatable test scenarios:
Test Case Definition
Test Case: Password Reset
Query: "How do I reset my password?"
Expected: Results should include password reset procedures
Success Criteria: Top 3 results contain reset instructions
Regression Testing
- Baseline Tests: Establish performance baselines
- Change Impact: Test after content updates
- Performance Regression: Monitor response time changes
- Quality Regression: Ensure result quality doesn't degrade
Continuous Testing
- Regular Test Runs: Schedule periodic testing
- Content Update Validation: Test after knowledge updates
- Performance Monitoring: Track search performance over time
- Quality Metrics: Monitor result relevance and accuracy
Test Result Analysis
Quality Metrics
Relevance Scoring
- Top Result Relevance: Is the first result what you need?
- Result Distribution: Are good results spread throughout or clustered?
- Irrelevant Results: How many results are not useful?
Completeness Assessment
- Information Coverage: Does the result contain complete information?
- Missing Information: What important details are missing?
- Context Adequacy: Is there enough context to understand the result?
Accuracy Validation
- Factual Correctness: Is the information accurate?
- Currency: Is the information up-to-date?
- Source Reliability: Are results from authoritative sources?
Performance Metrics
Response Time Analysis
- Average Response Time: Typical query response time
- Response Time Distribution: Variation in response times
- Slow Query Identification: Which queries are slowest?
- Performance Trends: How performance changes over time
Resource Usage
- Processing Load: Impact on system resources
- Concurrent User Impact: Performance with multiple users
- Scaling Behavior: Performance as knowledge base grows
Optimization Based on Testing
Content Optimization
Improve Poor Results
- Add Missing Content: Fill gaps identified during testing
- Improve Content Quality: Enhance poorly performing documents
- Update Outdated Information: Refresh stale content
- Remove Irrelevant Content: Clean up noise in results
Content Structure
- Better Organization: Improve document structure and headings
- Keyword Optimization: Ensure important terms are present
- Context Enhancement: Add context and background information
- Cross-References: Link related concepts and topics
Search Optimization
Parameter Tuning
- Result Limits: Adjust default number of results
- Relevance Thresholds: Fine-tune relevance scoring
- Search Algorithms: Optimize search parameters
- Chunking Strategy: Adjust content chunking for better results
Performance Tuning
- Index Optimization: Optimize search indexes
- Caching Strategy: Cache frequent queries
- Resource Allocation: Adjust system resources
- Concurrent Limits: Optimize for concurrent users
Integration Testing
Workflow Integration
Context Testing Test knowledge in actual workflows:
- Create Test Workflow: Simple workflow using knowledge
- Run with Test Data: Execute with various inputs
- Validate AI Responses: Check AI uses knowledge correctly
- Monitor Performance: Ensure workflow performance is acceptable
Real-World Scenarios
- Customer Support: Test with actual customer questions
- Document Processing: Test with real documents and tasks
- Decision Support: Test with actual business scenarios
- Content Generation: Test knowledge-based content creation
API Testing
Programmatic Access
- API Queries: Test knowledge retrieval via API
- Batch Testing: Test multiple queries efficiently
- Error Handling: Test API error responses
- Performance: Test API response times
Troubleshooting Test Issues
Common Problems
No Results Found
- Check Content Processing: Verify documents were indexed
- Review Query Terms: Try different keywords or phrasings
- Check Content Quality: Ensure source documents contain the information
- Validate Permissions: Confirm access to knowledge content
Poor Result Quality
- Content Issues: Review source document quality
- Processing Problems: Check text extraction quality
- Query Refinement: Try more specific or different queries
- Index Optimization: Consider rebuilding search indexes
Slow Performance
- Resource Issues: Check system resource availability
- Index Problems: Optimize or rebuild search indexes
- Query Complexity: Simplify complex queries
- Concurrent Load: Reduce concurrent query load
Debugging Steps
Result Analysis
- Review Raw Results: Check what content is actually returned
- Check Source Documents: Verify information exists in sources
- Analyze Processing Logs: Review content processing logs
- Test Alternative Queries: Try different ways to ask the same question
Performance Analysis
- Monitor System Resources: Check CPU, memory, and disk usage
- Analyze Query Patterns: Identify expensive or slow queries
- Review Index Status: Check search index health and optimization
- Test System Limits: Identify performance bottlenecks
Best Practices
Testing Strategy
- Regular Testing: Test knowledge bases regularly, not just after creation
- Comprehensive Coverage: Test all types of content and queries
- User Perspective: Test from the user's point of view
- Performance Baselines: Establish and maintain performance baselines
Quality Assurance
- Multiple Test Cases: Use diverse test scenarios
- Edge Case Testing: Test unusual or difficult queries
- Regression Testing: Ensure changes don't break existing functionality
- User Feedback: Incorporate actual user experiences
Documentation
- Test Results: Document test outcomes and optimizations
- Known Issues: Track and document known limitations
- Performance Metrics: Maintain performance history
- Optimization History: Record changes and their impacts
Thorough testing ensures your knowledge base provides reliable, accurate information for your AI workflows. Continue to Refresh Knowledge to learn about maintaining and updating your knowledge base over time.