Skip to content

Commit 1848ca0

Browse files
committed
reformat code
1 parent da2941f commit 1848ca0

File tree

13 files changed

+102
-80
lines changed

13 files changed

+102
-80
lines changed

paddleformers/transformers/deepseek_v2/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
# limitations under the License.
1515
""" DeepSeekV2 model configuration"""
1616
from ..configuration_utils import PretrainedConfig
17-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
17+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1818

1919
__all__ = [
2020
"DeepseekV2Config",

paddleformers/transformers/ernie4_5/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414

1515
""" Ernie4_5 model configuration."""
1616
from ..configuration_utils import PretrainedConfig
17-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
17+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1818

1919

2020
class Ernie4_5Config(PretrainedConfig):

paddleformers/transformers/ernie4_5_moe/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818

1919
from ...utils.log import logger
2020
from ..configuration_utils import PretrainedConfig
21-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
21+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
2222

2323
__all__ = ["Ernie4_5_MoeConfig"]
2424

paddleformers/transformers/glm4_moe/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
# limitations under the License.
1414

1515
from ..configuration_utils import PretrainedConfig
16-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
16+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1717

1818

1919
class Glm4MoeConfig(PretrainedConfig):

paddleformers/transformers/gpt_oss/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414

1515
# from ..configuration_utils import PretrainedConfig, layer_type_validation
1616
from ..configuration_utils import PretrainedConfig
17-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
17+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1818

1919

2020
class GptOssConfig(PretrainedConfig):

paddleformers/transformers/llama/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
""" Llama model configuration"""
1616

1717
from ..configuration_utils import PretrainedConfig
18-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
18+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1919

2020
__all__ = [
2121
"LLAMA_PRETRAINED_INIT_CONFIGURATION",

paddleformers/transformers/modeling_rope_utils.py

Lines changed: 13 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,14 @@
1313
# limitations under the License.
1414

1515
import math
16-
import paddle
1716
from typing import Optional
1817

18+
import paddle
19+
1920
from ..utils.log import logger
2021
from .configuration_utils import PretrainedConfig
2122

23+
2224
def standardize_rope_params(config, rope_theta: float | dict[str, float] | None = None):
2325
"""
2426
Helper to standardize the config's rope params field by ensuring the params are defined for each
@@ -69,6 +71,7 @@ def standardize_rope_params(config, rope_theta: float | dict[str, float] | None
6971
}
7072
config.rope_parameters = rope_parameters_per_layer_type
7173

74+
7275
def _compute_linear_scaling_rope_parameters(
7376
config: Optional[PretrainedConfig] = None,
7477
device: Optional[str] = None,
@@ -416,7 +419,7 @@ def _compute_longrope_parameters(
416419
else:
417420
ext_factors = paddle.to_tensor(short_factor, dtype=paddle.float32, place=device)
418421
inv_freq_shape = paddle.arange(0, dim, 2, dtype="int64", device=device).float() / dim
419-
inv_freq = 1.0 / (ext_factors * base ** inv_freq_shape)
422+
inv_freq = 1.0 / (ext_factors * base**inv_freq_shape)
420423

421424
return inv_freq, attention_factor
422425

@@ -511,6 +514,7 @@ def _compute_llama3_parameters(
511514
"llama3": _compute_llama3_parameters,
512515
}
513516

517+
514518
def _check_received_keys(
515519
rope_type: str,
516520
received_keys: set,
@@ -539,6 +543,7 @@ def _check_received_keys(
539543
if unused_keys:
540544
logger.warning(f"Unrecognized keys in `rope_parameters` for 'rope_type'='{rope_type}': {unused_keys}")
541545

546+
542547
def _validate_default_rope_parameters(
543548
rope_parameters: dict, config: Optional[PretrainedConfig] = None, ignore_keys: Optional[set] = None
544549
):
@@ -547,6 +552,7 @@ def _validate_default_rope_parameters(
547552
rope_type = rope_parameters["rope_type"]
548553
_check_received_keys(rope_type, received_keys, required_keys, ignore_keys=ignore_keys)
549554

555+
550556
def _validate_linear_scaling_rope_parameters(
551557
rope_parameters: dict, config: Optional[PretrainedConfig] = None, ignore_keys: Optional[set] = None
552558
):
@@ -559,6 +565,7 @@ def _validate_linear_scaling_rope_parameters(
559565
if factor is None or not isinstance(factor, float) or factor < 1.0:
560566
logger.warning(f"`rope_parameters`'s factor field must be a float >= 1, got {factor}")
561567

568+
562569
def _validate_dynamic_scaling_rope_parameters(
563570
rope_parameters: dict, config: Optional[PretrainedConfig] = None, ignore_keys: Optional[set] = None
564571
):
@@ -584,7 +591,7 @@ def _validate_yarn_parameters(
584591
"original_max_position_embeddings",
585592
"mscale",
586593
"mscale_all_dim",
587-
"truncate"
594+
"truncate",
588595
}
589596
received_keys = set(rope_parameters.keys())
590597
rope_type = rope_parameters["rope_type"]
@@ -729,6 +736,7 @@ def _validate_llama3_parameters(rope_parameters: dict, config: PretrainedConfig,
729736
f"{original_max_position_embeddings} and max_position_embeddings={config.max_position_embeddings}"
730737
)
731738

739+
732740
# Like `ROPE_INIT_FUNCTIONS`, this validation function mapping can be dynamically updated for custom RoPE types.
733741
ROPE_VALIDATION_FUNCTIONS = {
734742
"default": _validate_default_rope_parameters,
@@ -739,6 +747,7 @@ def _validate_llama3_parameters(rope_parameters: dict, config: PretrainedConfig,
739747
"llama3": _validate_llama3_parameters,
740748
}
741749

750+
742751
def rope_config_validation(config: PretrainedConfig, ignore_keys: Optional[set] = None):
743752
"""
744753
Validate the RoPE config arguments, given a `PreTrainedConfig` object
@@ -767,4 +776,4 @@ def rope_config_validation(config: PretrainedConfig, ignore_keys: Optional[set]
767776
else:
768777
logger.warning(
769778
f"Missing validation function mapping in `ROPE_VALIDATION_FUNCTIONS` for 'rope_type'='{rope_type}'"
770-
)
779+
)

paddleformers/transformers/qwen/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
# limitations under the License.
1414

1515
from ..configuration_utils import PretrainedConfig
16-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
16+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1717

1818
__all__ = ["QWenConfig"]
1919

paddleformers/transformers/qwen2/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
"""Qwen2 model configuration"""
1616

1717
from ..configuration_utils import PretrainedConfig, layer_type_validation
18-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
18+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1919

2020

2121
class Qwen2Config(PretrainedConfig):

paddleformers/transformers/qwen2_moe/configuration.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
""" Qwen2Moe model configuration"""
1616

1717
from ..configuration_utils import PretrainedConfig
18-
from ..modeling_rope_utils import standardize_rope_params, rope_config_validation
18+
from ..modeling_rope_utils import rope_config_validation, standardize_rope_params
1919

2020

2121
class Qwen2MoeConfig(PretrainedConfig):

0 commit comments

Comments
 (0)