|
27 | 27 |
|
28 | 28 | AWS_CREDENTIALS_DIR = '/etc/boto_cfg/' |
29 | 29 | AWS_ENV_CREDENTIALS_PROVIDER = 'com.amazonaws.auth.EnvironmentVariableCredentialsProvider' |
| 30 | +AWS_DEFAULT_CREDENTIALS_PROVIDER = 'com.amazonaws.auth.DefaultAWSCredentialsProviderChain' |
30 | 31 | GPU_POOLS_YAML_FILE_PATH = '/nail/srv/configs/gpu_pools.yaml' |
31 | 32 | DEFAULT_PAASTA_VOLUME_PATH = '/etc/paasta/volumes.json' |
32 | 33 | DEFAULT_SPARK_MESOS_SECRET_FILE = '/nail/etc/paasta_spark_secret' |
@@ -348,9 +349,10 @@ def _get_k8s_spark_env( |
348 | 349 | spark_env.update( |
349 | 350 | { |
350 | 351 | 'spark.kubernetes.authenticate.serviceAccountName': service_account_name, |
| 352 | + 'spark.kubernetes.authenticate.executor.serviceAccountName': service_account_name, |
351 | 353 | }, |
352 | 354 | ) |
353 | | - elif not include_self_managed_configs: |
| 355 | + if not include_self_managed_configs: |
354 | 356 | spark_env.update({ |
355 | 357 | 'spark.master': f'k8s://{k8s_server_address}', |
356 | 358 | }) |
@@ -1125,8 +1127,8 @@ def get_spark_conf( |
1125 | 1127 | spark_conf = {**(spark_opts_from_env or {}), **_filter_user_spark_opts(user_spark_opts)} |
1126 | 1128 | random_postfix = utils.get_random_string(4) |
1127 | 1129 |
|
1128 | | - if aws_creds is not None and aws_creds[2] is not None: |
1129 | | - spark_conf['spark.hadoop.fs.s3a.aws.credentials.provider'] = AWS_ENV_CREDENTIALS_PROVIDER |
| 1130 | + if (aws_creds is not None and aws_creds[2] is not None) or service_account_name is not None: |
| 1131 | + spark_conf['spark.hadoop.fs.s3a.aws.credentials.provider'] = AWS_DEFAULT_CREDENTIALS_PROVIDER |
1130 | 1132 |
|
1131 | 1133 | # app_name from env is already appended with port and time to make it unique |
1132 | 1134 | app_name = (spark_opts_from_env or {}).get('spark.app.name') |
|
0 commit comments