๐Ÿ” DPR: Semantic Search & Passage Retrieval Demo

How it works

This app uses Dense Passage Retrieval (DPR) โ€” a neural approach to information retrieval that understands meaning, not just keywords.

Unlike traditional keyword search (BM25 / TF-IDF), DPR encodes questions and passages into dense vector representations and finds the best match via dot-product similarity in embedding space.

1 14
๐Ÿ’ฌ Try these example questions

๐Ÿ”„ Dual-Encoder Workflow

Step Component What happens
1 Context Encoder Each passage in the knowledge base is encoded into a 768-dim vector (done once at startup)
2 Question Encoder Your query is encoded into a 768-dim vector (done at search time)
3 Dot Product We compute score = Q ยท Pแต€ between the query vector and every passage vector
4 Ranking Passages are ranked by score โ€” the highest score = most semantically relevant

๐Ÿ’ก Key insight: "Who landed on the Moon?" will match the Apollo 11 passage even though the passage never contains the word "landed on the Moon" verbatim โ€” because the meaning is captured in the vector space.

Models used: