add config examples to use PQ and SQ indexing and search for wiki-1M with Cohere embeddings#1047
add config examples to use PQ and SQ indexing and search for wiki-1M with Cohere embeddings#1047harsha-simhadri wants to merge 7 commits into
Conversation
There was a problem hiding this comment.
Pull request overview
Adds new benchmark configuration examples intended to demonstrate product quantization (PQ) and spherical quantization workflows for the wikipedia-1M Cohere embedding dataset.
Changes:
- Added a PQ graph-index build+search example JSON for wikipedia-1M.
- Added an exhaustive spherical-quantization example JSON (currently named as a wiki1M graph-index example).
Reviewed changes
Copilot reviewed 1 out of 2 changed files in this pull request and generated 1 comment.
| File | Description |
|---|---|
| diskann-benchmark/example/graph-index-spherical-quantization-wiki1M.json | Adds an exhaustive spherical quantization benchmark config (but currently references siftsmall test data / exhaustive tag). |
| diskann-benchmark/example/graph-index-product-quantization-wiki1M.json | Adds a PQ graph-index benchmark config for wikipedia-1M (currently uses an unregistered job type). |
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
Co-authored-by: Copilot Autofix powered by AI <175728472+Copilot@users.noreply.github.com>
The PR accidentally renamed spherical-exhaustive.json to graph-index-spherical-quantization-wiki1M.json, breaking the spherical_quantization_intergration test which references the old name. - Restore spherical-exhaustive.json with its original content - Update graph-index-spherical-quantization-wiki1M.json with proper content Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #1047 +/- ##
==========================================
- Coverage 90.60% 89.47% -1.13%
==========================================
Files 461 461
Lines 85494 85559 +65
==========================================
- Hits 77462 76558 -904
- Misses 8032 9001 +969
Flags with carried forward coverage won't be shown. Click here to find out more. 🚀 New features to boost your workflow:
|
| @@ -0,0 +1,55 @@ | |||
| { | |||
| "search_directories": [ | |||
| "../big-ann-benchmarks/data/wikipedia_cohere" ], | |||
There was a problem hiding this comment.
Minor: In the spirit of self‑service, consider including guidance on how to download the dataset, so that even an AI agent can follow the steps and retrieve the required data.
There was a problem hiding this comment.
where do you prefer I add this?
There was a problem hiding this comment.
added to benchmark readme. please mark as resolved if this is acceptable
hildebrandmw
left a comment
There was a problem hiding this comment.
Thanks. While a pre-built self serve like this is nice, I worry that if we proliferate these kinds of auxiliary files without a CI check for compatibility, they will continually break as we evolve diskann-benchmark. Is this just a convenience for not using the inputs functionality of diskann-benchmark and if so, is there something we can do the to CLI instead to make it more ergonomic?
| "source":{ | ||
| "index-source": "Build", | ||
| "data_type": "float32", | ||
| "data": "wikipedia_base.bin.crop_nb_1000000", |
There was a problem hiding this comment.
Would be nice if we could run these through the dry run functionality to ensure they don't get out of date. The one problem would be file-path validation would fail without the existence of the data/query/groundtruth files.
There was a problem hiding this comment.
I will look into dry run checks.
The goals of the new files is to illustrate the parameter choices to be used for larger datasets. The params in config files for test datasets are insufficient for larger datasets. If on the other hand you are oppposed to adding these entries here, please let me know and I can drop it, and not spend time on dry run checks
There was a problem hiding this comment.
Relatedly, it looks like there is no example config that builds a PQ index using the example datasets in this repo, although there is one for spherical. Is that something we should change?
There was a problem hiding this comment.
I think adding examples is great and would really help reduce friction for people trying out the library! My concern is around ensuring these examples are up-to-date, so in the future users don't run into the unfortunate situation where they try an example and it immediately fails for reasons other than the files not being in the correct place.
Apologies if my original comment didn't leave much direction. The main issue with the existing dry-run functionality is that file paths are also validated (at least for existence) as part of the dry-run check, which makes it harder to use it out-of-the-box.
The issue boils down to this check here. We want a way to turn it off when running a CI validation run, then we'd be good. The problem is how to turn it off cleanly in a way that is somewhat general. Maybe an additional "--no-environment" flag in addition to --dry-run to skip system dependent environment checks and making that available in the Checker struct? That would provide a nice place for generally disabling all kinds of system dependent checks for CI purposes.
add config examples to use PQ and SQ indexing and search for wiki-1M with Cohere embeddings