Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MaskFnAttentionBias._bool_value passes the same rank position tensors to mask_fn. #888

Merged
merged 1 commit into from
Dec 17, 2024

Conversation

ds-hwang
Copy link
Contributor

MaskFnAttentionBias._bool_value passes the same rank position tensors to mask_fn.

When target_positions is set, a rank 3 target_positions and a rank 2
source_positions are passed to mask_fn. From the perspective of a downstream
defining mask_fn, this is a big surprise.

@ds-hwang
Copy link
Contributor Author

From 950

@ruomingp ruomingp requested a review from apghml December 12, 2024 17:42
Copy link
Contributor

@apghml apghml left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This PR is under discussion internally.

@ds-hwang
Copy link
Contributor Author

Hi, could you take a look? From 950

Copy link
Contributor

@apghml apghml left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

… to mask_fn.

When target_positions is set, a rank 3 target_positions and a rank 2
source_positions are passed to mask_fn. From the perspective of a downstream
defining mask_fn, this is a big surprise.
@ds-hwang
Copy link
Contributor Author

Thank you for review!

@ds-hwang ds-hwang added this pull request to the merge queue Dec 17, 2024
Merged via the queue into apple:main with commit a7e2a95 Dec 17, 2024
6 checks passed
@ds-hwang ds-hwang deleted the bias_fix branch December 17, 2024 21:55
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants