Why I built this
Scientific QA needs citations and multi-paper synthesis. Most systems produce fluent answers with weak traceability, and I wanted to explore how to enforce citation discipline.
What I learned
Structured synthesis templates enforced citation discipline better than post-hoc verification. Multi-paper retrieval requires careful query expansion and reranking. What didn't work: initial over-reliance on single-paper retrieval.
TL;DR
Built a citation-grounded QA system that retrieves from multiple scientific papers, synthesizes answers with explicit citation anchors, and verifies citations post-generation.
Why this matters
Scientific QA needs citations and multi-paper synthesis. Most systems produce fluent answers with weak traceability.
Problem
Scientific QA needs citations and multi-paper synthesis. Most systems produce fluent answers with weak traceability.
Constraints
Multi-paper scope: answers must synthesize information across multiple research papers. Citation accuracy: every claim must map to specific paper sections. Latency: retrieval and synthesis must complete in reasonable time.
System design
Evaluation
Results
Stronger grounding and traceability through citation-first generation and verification gates.
Trade-offs & Lessons
Multi-paper retrieval increases compute but improves reliability. Structured synthesis templates enforced citation discipline. The initial over-reliance on single-paper retrieval was a mistake. Multi-paper retrieval requires careful query expansion and reranking.
What I'd Improve Next
Add citation quality scoring to prioritize high-impact papers. Implement cross-paper consistency checking. Build domain-specific retrieval strategies for different scientific fields.