generated from amazon-archives/__template_MIT-0
-
Notifications
You must be signed in to change notification settings - Fork 33
Closed
Labels
bugSomething isn't workingSomething isn't working
Description
Describe the bug
Flash attention integration tests failed
Expected Behavior
Flash attention integration correctness pytest should pass.
Current Behavior
FAILED flash_attention_correctness.py::test_attention[True-dtype0] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[True-dtype1] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[False-dtype0] - TypeError: multiple values for argument 'softmax_scale'
FAILED flash_attention_correctness.py::test_attention[False-dtype1] - TypeError: multiple values for argument 'softmax_scale'
Reproduction Steps
Running
pytest flash_attention_correctness.py
from test/integration/flash_attention.
Regression Issue
- Select this option if this issue appears to be a regression.
Possible Solution
It is currently calling compiler NKI kernel. Upgrade to the one from the repo?
Additional Information/Context
No response
neuronx-cc version used
2.21
Framework(s) and their versions used (JAX, PyTorch, etc..)
No response
shawnzxf
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working