Skip to content

Commit 05b044e

Browse files
punitvarahmellor
andauthored
[Doc] Fix cross-reference warnings (vllm-project#25058)
Signed-off-by: Punit Vara <[email protected]> Signed-off-by: Harry Mellor <[email protected]> Co-authored-by: Harry Mellor <[email protected]>
1 parent aa3f105 commit 05b044e

File tree

6 files changed

+18
-14
lines changed

6 files changed

+18
-14
lines changed

vllm/benchmarks/datasets.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -171,7 +171,8 @@ def get_random_lora_request(
171171
If `None`, LoRA is not used.
172172
173173
Returns:
174-
A new [LoRARequest][] (or `None` if not applicable).
174+
A new [`LoRARequest`][vllm.lora.request.LoRARequest]
175+
(or `None` if not applicable).
175176
"""
176177
if max_loras is None or lora_path is None:
177178
return None

vllm/distributed/device_communicators/shm_object_storage.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ class SingleWriterShmRingBuffer:
3030
- Maintains metadata for each allocated buffer chunk in the writer process
3131
- Supports custom "is_free_fn" functions to determine when buffers can be
3232
reused
33-
- Each buffer chunk contains: [4-byte id][4-byte size][actual_data]
33+
- Each buffer chunk contains: `[4-byte id][4-byte size][actual_data]`
3434
3535
Key Concepts:
3636
- monotonic_id_start/end: Track the range of active buffer IDs
@@ -99,7 +99,7 @@ class SingleWriterShmRingBuffer:
9999
- Writer handles garbage collection (free_buf) based on reader feedback
100100
101101
Memory Layout per Buffer Chunk:
102-
[4-byte monotonic_id][4-byte chunk_size][actual_data...]
102+
`[4-byte monotonic_id][4-byte chunk_size][actual_data...]`
103103
^metadata_start ^data_start
104104
105105
The monotonic_id ensures data integrity - readers can verify they're
@@ -185,7 +185,7 @@ def allocate_buf(self, size: int) -> tuple[int, int]:
185185
'''
186186
Allocate a buffer `MD_SIZE` + `size` bytes in the shared memory.
187187
Memory layout:
188-
[4-byte monotonic_id][4-byte size][buffer data...]
188+
`[4-byte monotonic_id][4-byte size][buffer data...]`
189189
'''
190190
assert self.is_writer, "Only the writer can allocate buffers."
191191
assert size > 0, "Size must be greater than 0"
@@ -413,7 +413,7 @@ class SingleWriterShmObjectStorage:
413413
allocation
414414
415415
Memory Layout per Object:
416-
[4-byte reference_count][metadata_size][serialized_object_data]
416+
`[4-byte reference_count][metadata_size][serialized_object_data]`
417417
418418
Thread Safety:
419419
- Writer operations (put, clear) are single-threaded by design

vllm/model_executor/layers/mamba/ops/causal_conv1d.py

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -927,11 +927,13 @@ def causal_conv1d_update(
927927
validate_data=False,
928928
):
929929
"""
930-
x: (batch, dim) or (batch, dim, seqlen) or (num_tokens, dim)
931-
[shape=2: single token prediction]
932-
[shape=3: single or multiple tokens prediction]
933-
[shape=2 with num_tokens: continuous batching, where num_tokens is the
934-
total tokens of all sequences in that batch]
930+
x: Input tensor which can take the following shapes:
931+
932+
- `[batch, dim]` - single token prediction
933+
- `[batch, dim, seqlen]` - single or multiple tokens prediction
934+
- `[num_tokens, dim]` - continuous batching, where num_tokens is
935+
the total tokens of all sequences in that batch
936+
935937
conv_state: (..., dim, state_len), where state_len >= width - 1
936938
weight: (dim, width)
937939
bias: (dim,)

vllm/model_executor/models/mistral3.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -583,7 +583,7 @@ def forward(
583583
inputs_embeds: Optional tensor of input embeddings.
584584
585585
Info:
586-
[Mistral3ImagePixelInputs][]
586+
[`Mistral3ImagePixelInputs`][vllm.model_executor.models.mistral3.Mistral3ImagePixelInputs]
587587
"""
588588
if intermediate_tensors is not None:
589589
inputs_embeds = None

vllm/multimodal/profiling.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -301,7 +301,7 @@ def get_mm_max_contiguous_tokens(
301301
Returns the maximum length of the multimodal (image placeholders+text)
302302
tokens, including any break/text tokens in-between image embeddings.
303303
304-
<im_start> [IMG] [IMG] [IMG] <row_break> [IMG] [IMG] [IMG] <im_end>
304+
`<im_start> [IMG] [IMG] [IMG] <row_break> [IMG] [IMG] [IMG] <im_end>`
305305
Returns 9, even when the number of image embeddings is 6.
306306
307307
This is important to take into account when profiling and

vllm/v1/core/kv_cache_manager.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,9 @@ class KVCacheBlocks:
2424
"""
2525
blocks: tuple[list[KVCacheBlock], ...]
2626
"""
27-
blocks[i][j] refers to the i-th kv_cache_group and the j-th block of tokens.
28-
We don't use block of tokens as the outer dimension because it assumes all
27+
`blocks[i][j]` refers to the i-th kv_cache_group
28+
and the j-th block of tokens.We don't use block of
29+
tokens as the outer dimension because it assumes all
2930
kv_cache_groups have the same number of blocks, which is true for now but
3031
will be broken if we want to give different block_size to different
3132
kv_cache_groups in the future.

0 commit comments

Comments
 (0)