LLM: relax batch check of flash atttention by double check attention mask #14251
Triggered via pull request
February 28, 2024 09:19
Status
Cancelled
Total duration
2m 17s
Artifacts
–
python-style-check.yml
on: pull_request
create-workflow-badge
0s
Matrix: style-check
Annotations
2 errors
style-check (3.7)
Canceling since a higher priority waiting request for 'Python Style Check-10270' exists
|
style-check (3.7)
The operation was canceled.
|