Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Moved FoldPrePostReshapeSoftmax into LLaMA-specific pattern matching #1122

Open
uranus0515 opened this issue Nov 9, 2023 · 0 comments
Open
Assignees
Labels
area-Compile bug Something isn't working

Comments

@uranus0515
Copy link
Contributor

Describe the bug
The FoldPrePostReshapeSoftmax is added for LLaMA,but the conditions of FoldPrePostReshapeSoftmax are not strict enough, it will cause damage to other scenes. PR #1121 deletes FoldPrePostReshapeSoftmax in the general Transform, which will cause LLaMA matching errors.

To Reproduce
the input of the first reshape is inconsistent with the output of the second reshape.

Expected behavior
Write this reshape into LLaMA's pattern matching instead of as a general transform.

Origin model and code
p.Add<Passes.Rules.Neutral.FoldPrePostReshapeSoftmax>();
This line has been deleted, and it should be written into pattern matching unique to LLaMA

Environment (please complete the following information):

  • OS: [Ubuntu20.04]
  • nncase version [commit id 3817072]
  • DL Framework [TensorFlow 2.5.0]

Additional context
Add any other context about the problem here.

@sunnycase sunnycase added bug Something isn't working area-Compile labels Nov 9, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area-Compile bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants