Skip to content

Commit 71b1c8b

Browse files
authored
[Chore]:Extract math and argparse utilities to separate modules (#27188)
Signed-off-by: Yeshwanth Surya <[email protected]> Signed-off-by: Yeshwanth N <[email protected]> Signed-off-by: yeshsurya <[email protected]>
1 parent 8fb7b2f commit 71b1c8b

File tree

125 files changed

+715
-639
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

125 files changed

+715
-639
lines changed

benchmarks/benchmark_block_pool.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
from benchmark_utils import TimeCollector
66
from tabulate import tabulate
77

8-
from vllm.utils import FlexibleArgumentParser
8+
from vllm.utils.argparse_utils import FlexibleArgumentParser
99
from vllm.v1.core.block_pool import BlockPool
1010

1111

benchmarks/benchmark_long_document_qa_throughput.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646

4747
from vllm import LLM, SamplingParams
4848
from vllm.engine.arg_utils import EngineArgs
49-
from vllm.utils import FlexibleArgumentParser
49+
from vllm.utils.argparse_utils import FlexibleArgumentParser
5050

5151

5252
def test_long_document_qa(llm=None, sampling_params=None, prompts=None):

benchmarks/benchmark_ngram_proposer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
VllmConfig,
2020
)
2121
from vllm.platforms import current_platform
22-
from vllm.utils import FlexibleArgumentParser
22+
from vllm.utils.argparse_utils import FlexibleArgumentParser
2323
from vllm.v1.spec_decode.ngram_proposer import NgramProposer
2424
from vllm.v1.worker.gpu_input_batch import InputBatch
2525
from vllm.v1.worker.gpu_model_runner import GPUModelRunner

benchmarks/benchmark_prefix_caching.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737

3838
from vllm import LLM, SamplingParams
3939
from vllm.engine.arg_utils import EngineArgs
40-
from vllm.utils import FlexibleArgumentParser
40+
from vllm.utils.argparse_utils import FlexibleArgumentParser
4141

4242
try:
4343
from vllm.transformers_utils.tokenizer import get_tokenizer

benchmarks/benchmark_prioritization.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
from transformers import AutoTokenizer, PreTrainedTokenizerBase
1212

1313
from vllm.engine.arg_utils import EngineArgs
14-
from vllm.utils import FlexibleArgumentParser
14+
from vllm.utils.argparse_utils import FlexibleArgumentParser
1515

1616

1717
# Select a equi-probable random priority

benchmarks/benchmark_serving_structured_output.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@
5151
from backend_request_func import get_tokenizer
5252

5353
try:
54-
from vllm.utils import FlexibleArgumentParser
54+
from vllm.utils.argparse_utils import FlexibleArgumentParser
5555
except ImportError:
5656
from argparse import ArgumentParser as FlexibleArgumentParser
5757

benchmarks/cutlass_benchmarks/sparse_benchmarks.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
from weight_shapes import WEIGHT_SHAPES
1616

1717
from vllm import _custom_ops as ops
18-
from vllm.utils import FlexibleArgumentParser
18+
from vllm.utils.argparse_utils import FlexibleArgumentParser
1919

2020
DEFAULT_MODELS = list(WEIGHT_SHAPES.keys())
2121
DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512]

benchmarks/cutlass_benchmarks/w8a8_benchmarks.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,8 @@
1818
from vllm.model_executor.layers.quantization.utils.fp8_utils import (
1919
w8a8_triton_block_scaled_mm,
2020
)
21-
from vllm.utils import FlexibleArgumentParser, cdiv
21+
from vllm.utils.argparse_utils import FlexibleArgumentParser
22+
from vllm.utils.math_utils import cdiv
2223

2324
DEFAULT_MODELS = list(WEIGHT_SHAPES.keys())
2425
DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512]

benchmarks/kernels/bench_per_token_quant_fp8.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010
from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8
1111
from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape
1212
from vllm.triton_utils import triton
13-
from vllm.utils import FlexibleArgumentParser
13+
from vllm.utils.argparse_utils import FlexibleArgumentParser
1414
from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE
1515

1616

benchmarks/kernels/benchmark_activation.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010
from vllm.model_executor.custom_op import CustomOp
1111
from vllm.platforms import current_platform
1212
from vllm.triton_utils import triton
13-
from vllm.utils import FlexibleArgumentParser
13+
from vllm.utils.argparse_utils import FlexibleArgumentParser
1414
from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE
1515

1616
batch_size_range = [1, 16, 32, 64, 128]

0 commit comments

Comments
 (0)