Skip to content

Conversation

@choijon5
Copy link
Contributor

@choijon5 choijon5 commented Oct 15, 2025

To allow users to easily try out Helion, I'm adding a Python notebook to run the softmax kernels. This notebook is linked from our documentation and our GitHub README to run it on Google Colab, which has NVIDIA T4 GPU for free. We achieve good speedup (1.8-1.9X) for this example on T4 and with autotune_effort="quick", it can be run in 5~6 mins on the free version of Colab, instead of 16-17 mins with autotune_effort="full".

Python notebook can be tested with pytest with nbmake. Added this notebook to be run on CI.

@choijon5 choijon5 requested a review from jansel October 15, 2025 07:46
@meta-cla meta-cla bot added the CLA Signed This label is managed by the Meta Open Source bot. label Oct 15, 2025
@choijon5 choijon5 requested a review from oulgen October 15, 2025 07:46
if [[ "${{ matrix.alias }}" == *"a10g"* ]]; then export HELION_DEV_LOW_VRAM=1; fi
# -rf: print failed tests
# --timeout: max allowed time for each test
pytest -rf --timeout=60
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

you need to limit this to test folder now

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

do you mean I should move this to the test folder?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

pytest -rf --timeout=60 test

"metadata": {},
"outputs": [],
"source": [
"!pip3 install torch==2.9.0 --index-url https://download.pytorch.org/whl/test/cu126\n",
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

collab env uses cuda12.6?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

it actually uses 12.4 but PyTorch no longer does 12.4 releases so 12.6 will do :)

User can easily try out Helion with this notebook from our documentation
website and GitHub readme. With autotune_effort=quick, it runs quickly
and shows good speedup (on T4 - 1.8X speedup).
…o run notebook on CI so that it can run the notebook cells to setup PyTorch and Helion.
@choijon5 choijon5 removed the request for review from jansel October 22, 2025 05:36
@choijon5 choijon5 merged commit 5b126e4 into main Oct 22, 2025
13 of 14 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants