docs: update embedding benchmarks (3.3.0)#525
docs: update embedding benchmarks (3.3.0)#525github-actions[bot] wants to merge 1 commit intomainfrom
Conversation
Greptile SummaryThis is an automated documentation PR that updates Key concerns:
Confidence Score: 1/5
Important Files Changed
Flowchart%%{init: {'theme': 'neutral'}}%%
flowchart TD
A[Release workflow triggers benchmark run] --> B[codegraph embed - index symbols]
B --> C{Symbols indexed > 0?}
C -- Yes --> D[Run embedding benchmarks per model]
D --> E[Record Hit@1/3/5/10, Misses, timing]
E --> F[Commit populated entry to EMBEDDING-BENCHMARKS.md]
C -- No or Error --> G[symbols: 0, models: empty]
G --> H[Commit empty entry as latest results - THIS PR]
H --> I[Latest results table renders blank]
F --> J[Latest results table shows valid benchmark data]
Last reviewed commit: "docs: update embeddi..." |
| **Version:** 3.3.0 | **Strategy:** structured | **Symbols:** 0 | **Date:** 2026-03-19 | ||
|
|
||
| | Model | Dim | Context | Hit@1 | Hit@3 | Hit@5 | Hit@10 | Misses | Embed | Search | | ||
| |-------|----:|--------:|------:|------:|------:|-------:|-------:|------:|-------:| | ||
|
|
||
| <!-- EMBEDDING_BENCHMARK_DATA | ||
| [ | ||
| { | ||
| "version": "3.3.0", | ||
| "date": "2026-03-19", | ||
| "strategy": "structured", | ||
| "symbols": 0, | ||
| "models": {} | ||
| }, |
There was a problem hiding this comment.
Empty benchmark result committed — symbols: 0, models: {}
The 3.3.0 entry records symbols: 0 and "models": {}, meaning no symbols were indexed and no model benchmarks were executed. The previous version (3.1.4) had 1,095 symbols across populated model results. An empty result like this indicates the workflow run #381 either failed during the benchmark phase (before any symbols were embedded) or completed a dry-run with no data, and the CI still committed these empty results as a valid benchmark update.
As it stands, the "Latest results" table renders completely empty and the symbols: 0 headline is misleading — it implies the codebase has no indexable symbols rather than reflecting a measurement failure. Merging this will overwrite a valid 1,095-symbol baseline with an empty entry as the authoritative "latest" result.
Suggested actions before merging:
- Verify why the workflow run produced 0 symbols (check run logs for errors in the embedding/indexing phase).
- If the run genuinely failed, close this PR and re-trigger the benchmark workflow once the issue is resolved.
- If
symbols: 0is intentional for some reason (e.g., a no-symbol build), add a comment in the JSON entry explaining it so future readers aren't confused.
There was a problem hiding this comment.
Agreed — this is a failed benchmark run. The PR should be closed and re-triggered after diagnosing why the workflow produced 0 symbols.
|
This PR should not be merged — the 3.3.0 embedding benchmark recorded 0 symbols and empty models, indicating the workflow run failed during the indexing phase. Merging would overwrite the valid 3.1.4 baseline (1,095 symbols) with empty data. Recommend closing and re-triggering the benchmark workflow once the issue is diagnosed. |
|
Root cause identified and fixed in #528: @huggingface/transformers is a devDependency, not a regular/optional dependency. In --npm mode, the benchmark installs codegraph into a temp dir where devDeps are absent. All 6 model workers crash on import, producing symbols: 0, models: {}. Fix: bench-config.js now explicitly installs @huggingface/transformers in the temp dir (matching the existing pattern for native platform packages). Also added a guard in update-embedding-report.js that rejects empty results instead of silently overwriting valid data. This PR should still be closed — once #528 merges, re-trigger the benchmark workflow to get valid 3.3.0 embedding data. |
Automated embedding benchmark update for 3.3.0 from workflow run #381.