Skip to content

Conversation

@kvenkateshan-meta
Copy link
Contributor

Summary:
Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call gpus-per-node in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

Changes here

  1. Introduced Slurm Version based GPU request configuration

  2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jul 23, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
@kvenkateshan-meta kvenkateshan-meta force-pushed the export-D78778304 branch 2 times, most recently from f67d4b6 to d06b519 Compare July 23, 2025 20:25
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:
Pull Request resolved: meta-pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 28, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 28, 2025
Summary:
Pull Request resolved: meta-pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 29, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
Summary:
Pull Request resolved: meta-pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

@facebook-github-bot facebook-github-bot merged commit ae55901 into meta-pytorch:main Jul 29, 2025
23 of 24 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants