Skip to content

Conversation

svmhdvn
Copy link
Contributor

@svmhdvn svmhdvn commented Sep 5, 2025

On busybox-based systems like Alpine Linux, wget does not have certain CLI flags such as '--no-config'. Thus, search for the existence of 'curl' first in the PATH before wget. wget2 is still the preferred download tool.

On busybox-based systems like Alpine Linux, wget does not have
certain CLI flags such as '--no-config'. Thus, search for the
existence of 'curl' first in the PATH before wget. wget2 is
still the preferred download tool.
@danbev danbev merged commit edea8a9 into ggml-org:master Sep 8, 2025
55 of 59 checks passed
bygreencn added a commit to bygreencn/whisper.cpp that referenced this pull request Sep 24, 2025
* ggerganov/master: (92 commits)
  ggml : Fix MKL detection by quoting BLAS_INCLUDE_DIRS (ggml-org#3426)
  whisper : prefer curl over wget in download scripts (ggml-org#3409)
  ci : remove brew installation of cmake for macos-latest (ggml-org#3408)
  tests : use CMake definitions for model/sample paths (ggml-org#3406)
  Handle negative value in padding (ggml-org#3389)
  models :  update`./models/download-ggml-model.cmd` to allow for tdrz download (ggml-org#3381)
  talk-llama : sync llama.cpp
  sync : ggml
  ggml: Add initial WebGPU backend (llama/14521)
  ggml : initial zDNN backend (llama/14975)
  common : handle mxfp4 enum
  ggml-quants : fix make_qp_quants NANs and IQ1 assertion errors (llama/15379)
  vulkan: disable spirv-opt for bfloat16 shaders (llama/15352)
  vulkan: Use larger workgroups for mul_mat_vec when M is small (llama/15355)
  vulkan: support sqrt (llama/15370)
  vulkan: Optimize argsort (llama/15354)
  vulkan: fuse adds (llama/15252)
  vulkan: Support mul_mat_id with f32 accumulators (llama/15337)
  vulkan: Add missing bounds checking to scalar/coopmat1 mul_mat_id (llama/15334)
  OpenCL: add initial FA support (llama/14987)
  ...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants