Basic Functionality Verification Items: End-to-End Test Cases from Knowledge Base Upload to Retrieval to Conversation
The value of end-to-end testing lies in confirming that the system is not just “a certain page can be opened” but that the entire chain from document ingestion to answer output is functional.
This topic has relatively clear public support. Dify officially provides a knowledge retrieval test entry point, Question Classifier node documentation, and knowledge base related documentation. Public articles also include practices around RAG retrieval, automated evaluation, and knowledge pipelines. Therefore, this can be written as a training piece on “the minimum reproducible test set for delivery sites.”
1. End-to-End Test Framework Confirmed by Public Sources
1. Knowledge Base Retrieval Testing Is Already an Officially Published Capability
Dify officially provides Knowledge Test Retrieval capabilities, indicating that “verify retrieval after upload, then verify Q&A” is a product-endorsed testing path.
2. End-to-End Verification Should Cover at Least Three Stages
Based on a synthesis of public sources, the minimum chain should cover:
- Document enters the knowledge base
- Whether retrieval produces a hit
- Whether the final application answer is generated based on the retrieved content
3. Extended Test Cases Should Cover Complex Documents and Error Queries
Public RAG articles repeatedly emphasize that complex PDFs, parameter tuning, and error query handling significantly impact the experience. These should all be included in delivery testing.
2. Recommended Minimum Test Cases
- Upload a document
- Wait for indexing to complete
- Create an application bound to the knowledge base
- Submit a question that should definitely produce a hit
- Observe whether the answer correctly cites the relevant content
3. Extended Test Cases
- Upload multiple documents
- Upload a PDF containing tables
- Adjust Top-K and Rerank settings
- Verify whether the system appropriately declines to answer incorrect questions
4. Delivery Recommendations
During training, provide partners with a standard set of test texts and expected results for quick on-site validation.
Public Source References
note.com
- “Information that should be there can’t be found” – Pitfalls encountered in Dify RAG chatbot development and building an automated evaluation system | https://note.com/kadinche/n/n87b77918dab9
- AI that “re-searches on its own”: The forefront of advanced RAG construction with DeepSeek-R1 and Dify | https://note.com/nocode_solutions/n/nbe6c159a5460
zenn.dev / Official Documentation / Other Public Pages
- Knowledge Retrieval Test | https://docs.dify.ai/ja/use-dify/knowledge/test-retrieval
- Question Classifier - Dify Docs | https://docs.dify.ai/ja/use-dify/nodes/question-classifier
- [Dify] RAG Complete Guide: Thorough Explanation of Mechanisms and Settings | https://zenn.dev/upgradetech/articles/ac9099a6489abe
Confirmed Information from Public Sources
- The official knowledge retrieval test capability is suitable as a minimum verification step before acceptance
- End-to-end testing should cover at least the three-stage chain of “upload -> retrieval -> conversation”
- Complex PDFs, parameter tuning, and error query handling should be included as extended test items