ragflow-optimizer: Auto-tune chunking parameters with LLM-based evaluation #13099
stranger00135
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
|
This ragflow-optimizer for auto-tuning chunking is brilliant! Why this matters:
Key innovations:
params = {
"chunk_size": [256, 512, 1024],
"overlap": [0, 50, 100],
"method": ["sentence", "token", "semantic"]
}
Extensions I'd love:
Integration idea: # Auto-tune on new doc types
optimizer.tune(
docs=new_doc_set,
eval_queries=sample_queries,
budget=100 # Evaluation calls
)We build adaptive RAG at RevolutionAI. This is exactly what production systems need! Have you tested on domain-specific corpora (legal, medical, code)? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi RAGFlow community! 👋
I built an open-source tool that automatically finds the best chunking parameters for your RAGFlow knowledge bases.
The Problem
Chunking parameters (chunk_size, overlap, auto_questions, etc.) have a huge impact on retrieval quality, but most of us pick them by feel. There's no easy way to systematically compare configurations.
The Solution
ragflow-optimizer runs automated experiments on your actual documents:
Different document types get independently optimized — HR policies might need different chunking than technical SOPs.
Quick Start
Works with OpenAI, DeepSeek, and DashScope for evaluation.
Links
Would love feedback from the community! What parameters do you find hardest to tune? What metrics matter most to you?
⭐ If you find it useful, a star would really help!
Beta Was this translation helpful? Give feedback.
All reactions