- 
                Notifications
    You must be signed in to change notification settings 
- Fork 58
Adding demo for running softmax kernel on Google colab #944
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
9aab2bf    to
    3b6a570      
    Compare
  
    | if [[ "${{ matrix.alias }}" == *"a10g"* ]]; then export HELION_DEV_LOW_VRAM=1; fi | ||
| # -rf: print failed tests | ||
| # --timeout: max allowed time for each test | ||
| pytest -rf --timeout=60 | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you need to limit this to test folder now
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
do you mean I should move this to the test folder?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
pytest -rf --timeout=60 test
        
          
                notebooks/softmax.ipynb
              
                Outdated
          
        
      | "metadata": {}, | ||
| "outputs": [], | ||
| "source": [ | ||
| "!pip3 install torch==2.9.0 --index-url https://download.pytorch.org/whl/test/cu126\n", | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
collab env uses cuda12.6?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it actually uses 12.4 but PyTorch no longer does 12.4 releases so 12.6 will do :)
User can easily try out Helion with this notebook from our documentation website and GitHub readme. With autotune_effort=quick, it runs quickly and shows good speedup (on T4 - 1.8X speedup).
…ebook with pytest.
…o run notebook on CI so that it can run the notebook cells to setup PyTorch and Helion.
3b6a570    to
    961c12b      
    Compare
  
    fe44786    to
    ea3e3ae      
    Compare
  
    
To allow users to easily try out Helion, I'm adding a Python notebook to run the softmax kernels. This notebook is linked from our documentation and our GitHub README to run it on Google Colab, which has NVIDIA T4 GPU for free. We achieve good speedup (1.8-1.9X) for this example on T4 and with autotune_effort="quick", it can be run in 5~6 mins on the free version of Colab, instead of 16-17 mins with autotune_effort="full".
Python notebook can be tested with pytest with nbmake. Added this notebook to be run on CI.