Predictive / AI-Driven Analytics

Vector Databases & Retrieval-Augmented Models

Check your understanding of retrieval building blocks behind modern RAG systems. From indexing to reranking, verify you can tune quality, speed, and cost.

In a vector database, the embedding dimension must ______.

match the index’s expected dimensionality

be prime

equal the number of documents

change at query time

Indexes are built for a fixed vector size. Mismatched dimensions cause errors or incorrect retrieval.

HNSW is a popular ANN index because it provides ______.

high recall with sublinear query time via navigable small-world graphs

only CPU-bound training

exact k-NN with linear scans

lossless compression of text

Hierarchical navigable graphs speed up nearest-neighbor search while maintaining strong recall at practical latencies.

Hybrid retrieval that combines BM25 and vector similarity often ______.

improves recall and grounding over either method alone

requires identical embeddings and tokens

breaks tokenization entirely

prevents use of filters

Lexical and semantic signals are complementary; hybrid strategies recover both exact terms and paraphrased meaning.

Reranking top-k candidates with a cross-encoder can ______.

boost precision by scoring query–document pairs more deeply

lower latency in all cases

eliminate hallucinations entirely

remove the need for chunking

Cross-encoders compute richer relevance scores. They add compute cost, so systems balance speed vs. quality.

Chunking with moderate overlap helps because it ______.

preserves context that spans chunk boundaries

forces shorter indexes

improves OCR quality automatically

guarantees no duplicates

Overlaps reduce context fragmentation, improving retrieval for concepts that span adjacent passages.

For cost-efficient storage at scale, systems often use ______.

product quantization or scalar quantization on vectors

sorting chunks alphabetically

gzip on token IDs only

PNG compression on embeddings

Quantization compresses dense vectors with controllable accuracy–memory trade-offs, enabling larger indexes.

Maximal Marginal Relevance (MMR) is used to ______.

force chronological ordering

select diverse yet relevant results to reduce redundancy

learn embeddings end-to-end

deduplicate documents offline only

MMR balances similarity to the query with dissimilarity to already chosen items, increasing coverage of facets.

Metadata filters (e.g., by source or date) are valuable in RAG because they ______.

remove the need for evaluation

replace vector search entirely

guarantee perfect answers

constrain retrieval to trusted, timely content

Filters enforce policy and freshness while keeping semantic search benefits. They are complementary to embeddings.

A good grounding practice is to store ______ with each chunk.

source identifiers and offsets for citation

GPU model name

only the cosine score

window size used at index time

Keeping source pointers enables traceable answers and post-hoc audits, improving trust and usability.

To reduce embedding costs in production, teams commonly ______.

re-embed on every query

push embeddings into cookies

cache vectors and avoid re-embedding unchanged text

compute vectors with random seeds only

Caching prevents duplicate work for repeated content or frequent queries, saving time and money.

Starter

Great beginning—keep exploring the core ideas and key trade-offs.

Solid

Strong grasp—practice applying these choices to real data and workloads.

Expert!

Excellent—your decisions reflect production-grade mastery.

What's your reaction?

Related Quizzes

1 of 9

Leave A Reply

Your email address will not be published. Required fields are marked *