Skip to content

Conversation

ggerganov
Copy link
Member

@ggerganov ggerganov commented Aug 10, 2025

There has been an increased traffic to http://whisper.ggerganov.com over the past few months and it's getting a bit expensive to host the whisper models on my Linode:

June:

image

July + Aug:

image

This PR changes the URLs in the examples to point to the Hugging Face models in order to redirect most of the traffic:

https://huggingface.co/ggerganov/whisper.cpp

I haven't tested if this actually works.

@ggerganov ggerganov requested a review from danbev August 10, 2025 08:50
Copy link
Member

@danbev danbev left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Tested locally and works as expected:

loadRemote: "https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-tiny.en.bin" stored in the IndexedDB

@ggerganov ggerganov merged commit b02242d into master Aug 10, 2025
55 checks passed
@ggerganov ggerganov deleted the gg/switch-ggml-model-host-to-hf branch August 10, 2025 10:00
bygreencn added a commit to bygreencn/whisper.cpp that referenced this pull request Sep 24, 2025
* ggerganov/master: (72 commits)
  node : add win platform check for require path (ggml-org#3363)
  ci : update main-cuda.Dockerfile (ggml-org#3371)
  whisper : fixed crash in GPU device selection on multi-GPU systems (ggml-org#3372)
  wasm : change ggml model host to HF (ggml-org#3369)
  ruby : Add ruby binding for max_len (ggml-org#3365)
  stream.wasm : add language selection support (ggml-org#3354)
  whisper : reset conv scheduler when CoreML is used (ggml-org#3350)
  ggml : remove old kompute, cann (skip) (ggml-org#3349)
  talk-llama : sync llama.cpp
  sync : ggml
  vulkan : add fp16 support for the conv_2d kernel (llama/14872)
  vulkan: skip empty set_rows to avoid invalid API usage (llama/14860)
  HIP: Enable Matrix cores for MMQ Kernels, Enable stream-K for CDNA 3 (llama/14624)
  CANN: Implement GLU ops (llama/14884)
  musa: fix build warnings (unused variable) (llama/14869)
  ggml-cpu : disable GGML_NNPA by default due to instability (llama/14880)
  metal: SSM_SCAN performance (llama/14743)
  opencl: add fused `rms_norm_mul` (llama/14841)
  ggml : remove invalid portPos specifiers from dot files (llama/14838)
  rpc : check for null buffers in get/set/copy tensor endpoints (llama/14868)
  ...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants