Skip to content

[Misc] Bump opencv-python dependecy version to 4.13#32668

Merged
Isotr0py merged 6 commits intovllm-project:mainfrom
Isotr0py:bump-opencv-version
Jan 22, 2026
Merged

[Misc] Bump opencv-python dependecy version to 4.13#32668
Isotr0py merged 6 commits intovllm-project:mainfrom
Isotr0py:bump-opencv-version

Conversation

@Isotr0py
Copy link
Copy Markdown
Member

@Isotr0py Isotr0py commented Jan 20, 2026

Purpose

  • Update opencv-python version in dependencies to get up-to-date security fix.

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

@mergify mergify bot added the ci/build label Jan 20, 2026
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly updates the opencv-python-headless dependency to version 4.13.0 across all relevant requirement files, addressing the stated purpose of applying security fixes. However, an accompanying change to the genai_perf dependency in requirements/nightly_torch_test.txt introduces an inconsistency with requirements/test.in. I've left a comment with a recommendation to resolve this to ensure consistent testing environments.

Copy link
Copy Markdown

@cursor cursor bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cursor Bugbot has reviewed your changes and found 1 potential issue.

Bugbot Autofix is OFF. To automatically fix reported issues with Cloud Agents, enable Autofix in the Cursor dashboard.

Comment @cursor review or bugbot run to trigger another review on this PR

@Isotr0py Isotr0py added the ready ONLY add when PR is ready to merge/full CI is needed label Jan 20, 2026
@Isotr0py
Copy link
Copy Markdown
Member Author

/gemini review

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) January 20, 2026 13:50
@Isotr0py Isotr0py disabled auto-merge January 21, 2026 17:01
@Isotr0py
Copy link
Copy Markdown
Member Author

Isotr0py commented Jan 21, 2026

Wait, seems this PR breaks the video loader, will take a further look tomorrow morning.

NVM, it should be issue from #32746 actually.

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
@Isotr0py Isotr0py force-pushed the bump-opencv-version branch from 31e6ec8 to 0ca0ffa Compare January 21, 2026 18:33
@DarkLight1337
Copy link
Copy Markdown
Member

It is still failing

@Isotr0py
Copy link
Copy Markdown
Member Author

It is still failing

Hmmm, it's weird that I can't reproduce the failure locally by running the single test. Let me double check it by running the full pipeline...

Log
tests/entrypoints/openai/test_translation_validation.py::test_audio_with_max_tokens[/home/mozf/LLM/gemma-3n-E2B-it] Launching RemoteOpenAIServer with: vllm serve /home/mozf/LLM/gemma-3n-E2B-it --enforce-eager --port 41795 --seed 0
(APIServer pid=1133848) INFO 01-22 17:14:11 [api_server.py:872] vLLM API server version 0.14.0rc1.dev434+g771c07882.d20260111
(APIServer pid=1133848) INFO 01-22 17:14:11 [utils.py:267] non-default args: {'model_tag': '/home/mozf/LLM/gemma-3n-E2B-it', 'port': 41795, 'model': '/home/mozf/LLM/gemma-3n-E2B-it', 'enforce_eager': True}
(APIServer pid=1133848) INFO 01-22 17:14:11 [model.py:541] Resolved architecture: Gemma3nForConditionalGeneration
(APIServer pid=1133848) INFO 01-22 17:14:11 [model.py:1559] Using max model len 32768
(APIServer pid=1133848) INFO 01-22 17:14:11 [scheduler.py:229] Chunked prefill is enabled with max_num_batched_tokens=2048.
(APIServer pid=1133848) INFO 01-22 17:14:11 [vllm.py:618] Asynchronous scheduling is enabled.
(APIServer pid=1133848) WARNING 01-22 17:14:11 [vllm.py:656] Enforce eager set, overriding optimization level to -O0
(APIServer pid=1133848) INFO 01-22 17:14:11 [vllm.py:756] Cudagraph is disabled under eager mode
/home/mozf/develop-projects/vllm/.venv/lib/python3.12/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
  import pynvml  # type: ignore[import]
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:20 [core.py:96] Initializing a V1 LLM engine (v0.14.0rc1.dev434+g771c07882.d20260111) with config: model='/home/mozf/LLM/gemma-3n-E2B-it', speculative_config=None, tokenizer='/home/mozf/LLM/gemma-3n-E2B-it', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=32768, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=/home/mozf/LLM/gemma-3n-E2B-it, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.NONE: 0>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['all'], 'splitting_ops': [], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [2048], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.NONE: 0>, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': False, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 0, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:23 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://198.18.0.1:53015 backend=nccl
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:23 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:33 [gpu_model_runner.py:3820] Starting to load model /home/mozf/LLM/gemma-3n-E2B-it...
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:33 [vllm.py:618] Asynchronous scheduling is enabled.
(EngineCore_DP0 pid=1134167) WARNING 01-22 17:14:33 [vllm.py:663] Inductor compilation was disabled by user settings, optimizations settings that are only active during inductor compilation will be ignored.
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:33 [vllm.py:756] Cudagraph is disabled under eager mode
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:34 [cuda.py:351] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
Loading safetensors checkpoint shards:   0% Completed | 0/3 [00:00<?, ?it/s]
Loading safetensors checkpoint shards:  33% Completed | 1/3 [00:00<00:01,  1.70it/s]
Loading safetensors checkpoint shards:  67% Completed | 2/3 [00:01<00:00,  1.16it/s]
Loading safetensors checkpoint shards: 100% Completed | 3/3 [00:02<00:00,  1.18it/s]
Loading safetensors checkpoint shards: 100% Completed | 3/3 [00:02<00:00,  1.21it/s]
(EngineCore_DP0 pid=1134167) 
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:36 [default_loader.py:291] Loading weights took 2.61 seconds
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:37 [gpu_model_runner.py:3917] Model loading took 10.2 GiB memory and 3.157163 seconds
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:37 [gpu_model_runner.py:4735] Encoder cache will be initialized with a budget of 2048 tokens, and profiled with 8 image items of the maximum feature size.
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:39 [gpu_worker.py:355] Available KV cache memory: 8.53 GiB
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:39 [kv_cache_utils.py:1307] GPU KV cache size: 223,728 tokens
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:39 [kv_cache_utils.py:1312] Maximum concurrency for 32,768 tokens per request: 25.97x
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:39 [core.py:272] init engine (profile, create kv cache, warmup model) took 2.01 seconds
(EngineCore_DP0 pid=1134167) WARNING 01-22 17:14:40 [vllm.py:663] Inductor compilation was disabled by user settings, optimizations settings that are only active during inductor compilation will be ignored.
(EngineCore_DP0 pid=1134167) INFO 01-22 17:14:40 [vllm.py:756] Cudagraph is disabled under eager mode
(APIServer pid=1133848) INFO 01-22 17:14:40 [api_server.py:663] Supported tasks: ['generate', 'transcription']
(APIServer pid=1133848) WARNING 01-22 17:14:40 [model.py:1372] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`.
(APIServer pid=1133848) INFO 01-22 17:14:40 [serving.py:227] Using default chat sampling params from model: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:40 [serving.py:149] Using default chat sampling params from model: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:40 [serving.py:185] Warming up chat template processing...
(APIServer pid=1133848) INFO 01-22 17:14:43 [chat_utils.py:599] Detected the chat template content format to be 'openai'. You can set `--chat-template-content-format` to override this.
(APIServer pid=1133848) INFO 01-22 17:14:43 [serving.py:221] Chat template warmup completed in 2807.7ms
(APIServer pid=1133848) INFO 01-22 17:14:43 [serving.py:80] Using default completion sampling params from model: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:43 [speech_to_text.py:113] Overwriting default completion sampling param with: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:43 [speech_to_text.py:140] Warming up audio preprocessing libraries...
(APIServer pid=1133848) INFO 01-22 17:14:46 [speech_to_text.py:176] Audio preprocessing warmup completed in 3.36s
(APIServer pid=1133848) INFO 01-22 17:14:46 [speech_to_text.py:203] Warming up multimodal input processor...
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:236] Input processor warmup completed in 9.75s
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:113] Overwriting default completion sampling param with: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:140] Warming up audio preprocessing libraries...
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:176] Audio preprocessing warmup completed in 0.00s
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:203] Warming up multimodal input processor...
(APIServer pid=1133848) INFO 01-22 17:14:56 [speech_to_text.py:236] Input processor warmup completed in 0.00s
(APIServer pid=1133848) INFO 01-22 17:14:56 [serving.py:149] Using default chat sampling params from model: {'top_k': 64, 'top_p': 0.95}
(APIServer pid=1133848) INFO 01-22 17:14:56 [api_server.py:946] Starting vLLM API server 0 on http://0.0.0.0:41795
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:38] Available routes are:
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /docs, Methods: GET, HEAD
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /redoc, Methods: GET, HEAD
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /tokenize, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /detokenize, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /inference/v1/generate, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /pause, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /resume, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /is_paused, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /metrics, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /health, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/chat/completions, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/chat/completions/render, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/responses, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/audio/translations, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/completions, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/completions/render, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/messages, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/models, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /load, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /version, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /ping, Methods: GET
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /ping, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /invocations, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /classify, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/embeddings, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /score, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/score, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /rerank, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v1/rerank, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /v2/rerank, Methods: POST
(APIServer pid=1133848) INFO 01-22 17:14:56 [launcher.py:46] Route: /pooling, Methods: POST
(APIServer pid=1133848) INFO:     Started server process [1133848]
(APIServer pid=1133848) INFO:     Waiting for application startup.
(APIServer pid=1133848) INFO:     Application startup complete.
(APIServer pid=1133848) INFO:     127.0.0.1:39244 - "GET /health HTTP/1.1" 200 OK
(APIServer pid=1133848) INFO:     127.0.0.1:39250 - "POST /v1/audio/translations HTTP/1.1" 200 OK
I
(APIServer pid=1133848) INFO:     127.0.0.1:39262 - "POST /v1/audio/transcriptions HTTP/1.1" 200 OK
Hey, first words I'm speaking in the original corner. Uh little speech. Boy. No, and that way. No, and that way. No, and that way. No, and that way.
PASSED(APIServer pid=1133848) INFO 01-22 17:15:01 [launcher.py:110] Shutting down FastAPI HTTP server.
[rank0]:[W122 17:15:01.811289454 ProcessGroupNCCL.cpp:1524] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
(APIServer pid=1133848) INFO:     Shutting down
(APIServer pid=1133848) INFO:     Waiting for application shutdown.
(APIServer pid=1133848) INFO:     Application shutdown complete.
(APIServer pid=1133848) 

========================================================================================== warnings summary ==========================================================================================
.venv/lib/python3.12/site-packages/torch/cuda/__init__.py:63
  /home/mozf/develop-projects/vllm/.venv/lib/python3.12/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
    import pynvml  # type: ignore[import]

<frozen importlib._bootstrap>:488
  <frozen importlib._bootstrap>:488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute

<frozen importlib._bootstrap>:488
  <frozen importlib._bootstrap>:488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute

.venv/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305
  /home/mozf/develop-projects/vllm/.venv/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable.
    ref_error: type[Exception] = jsonschema.RefResolutionError,

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
============================================================================== 1 passed, 4 warnings in 61.93s (0:01:01) ==============================================================================

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
@Isotr0py Isotr0py enabled auto-merge (squash) January 22, 2026 15:18
@Isotr0py Isotr0py merged commit 444e2e7 into vllm-project:main Jan 22, 2026
96 checks passed
@Isotr0py Isotr0py deleted the bump-opencv-version branch January 22, 2026 15:51
khluu pushed a commit that referenced this pull request Jan 23, 2026
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
(cherry picked from commit 444e2e7)
monajafi-amd pushed a commit to monajafi-amd/vllm that referenced this pull request Jan 23, 2026
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: mohammad najafi <mohammad.najafi@amd.com>
cwazai pushed a commit to cwazai/vllm that referenced this pull request Jan 25, 2026
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: 陈建华 <1647430658@qq.com>
lapy pushed a commit to lapy/vllm that referenced this pull request Jan 27, 2026
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
yma11 pushed a commit to yma11/vllm that referenced this pull request Feb 3, 2026
* Replace urllib's `urlparse` with urllib3's `parse_url` (vllm-project#32746)

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
(cherry picked from commit 8ebf271)

* Bump opencv-python dependecy version to 4.13 (vllm-project#32668)

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
(cherry picked from commit 444e2e7)

* Fix Whisper/encoder-decoder GPU memory leak (vllm-project#32789)

Signed-off-by: NickLucche <nlucches@redhat.com>
(cherry picked from commit ea6102b)

---------

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: NickLucche <nlucches@redhat.com>
Co-authored-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Co-authored-by: Nicolò Lucchesi <nlucches@redhat.com>
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
dtrifiro pushed a commit to dtrifiro/vllm that referenced this pull request Mar 9, 2026
…t#32668)

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
Signed-off-by: Daniele Trifirò <dtrifiro@redhat.com>
dtrifiro added a commit to dtrifiro/vllm that referenced this pull request Mar 9, 2026
Fix:
GHSA-4r2x-xpjr-7cvv

- [Frontend] Improve error message (vllm-project#31987)
- [Frontend] Standardize use of `create_error_response` (vllm-project#32319)
- [Misc] Bump opencv-python dependecy version to 4.13 (vllm-project#32668)

https://issues.redhat.com/browse/INFERENG-4411
dtrifiro added a commit to dtrifiro/vllm that referenced this pull request Mar 9, 2026
- [build] fix cu130 related release pipeline steps and publish as
nightly image (vllm-project#32522)
- [Misc] Replace urllib's `urlparse` with urllib3's `parse_url`
(vllm-project#32746)
- [Misc] Bump opencv-python dependency version to 4.13
(vllm-project#32668)
- [Bugfix] Fix Whisper/encoder-decoder GPU memory leak
(vllm-project#32789)
- [CI] fix version comparsion and exclusion patterns in
upload-release-wheels.sh (vllm-project#32971)
- tokenizers: mistral: fix merge conflict
- `Dockerfile.tpu.ubi`: add `git` to allow `pip install git+https`
Shafi-Hussain pushed a commit to odh-on-pz/vllm-cpu that referenced this pull request Mar 17, 2026
Fix:
GHSA-4r2x-xpjr-7cvv

- [Frontend] Improve error message (vllm-project/vllm#31987)
- [Frontend] Standardize use of `create_error_response` (vllm-project/vllm#32319)
- [Misc] Bump opencv-python dependecy version to 4.13 (vllm-project/vllm#32668)

https://issues.redhat.com/browse/INFERENG-4411
https://issues.redhat.com/browse/RHOAIENG-48365
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants