Skip to content

Allow TMA benchmarks for flex-attention kernel #225

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

mandroid6
Copy link
Contributor

Summary:
This diff adds a new argument --use-tma to the operator.py file in the flex_attention directory of the tritonbench repository. This argument allows users to enable Tensor Memory Access (TMA) in kernel options for flex-attention benchmarks.

Changes:

  • Added --use-tma argument to the parse_args function in operator.py
  • Modified the parse_args function to store the --use-tma value in the args object

Differential Revision: D74839480

Summary:
This diff adds a new argument `--use-tma` to the `operator.py` file in the `flex_attention` directory of the `tritonbench` repository. This argument allows users to enable Tensor Memory Access (TMA) in kernel options for flex-attention benchmarks.

**Changes:**

* Added `--use-tma` argument to the `parse_args` function in `operator.py`
* Modified the `parse_args` function to store the `--use-tma` value in the `args` object

Differential Revision: D74839480
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D74839480

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants